Learning and Motivation 33 (2002) 485–509 www.academicpress.com Between-subject PREE and within-subject reversed PREE in spaced-trial extinction with pigeons Mauricio R. Papini,* Brian L. Thomas, and Dawn G. McVicar Department of Psychology, Texas Christian University, Box 298920, Fort Worth, TX 76129, USA Received 13 December 2001; revised 19 March 2002 Abstract Three experiments explored the partial reinforcement extinction effect (PREE; greater resistance to extinction after partial, rather than continuous, reinforcement training), in a spaced-trial situation with pigeons. Experiments 1 and 2 report conventional PREEs with 24-h intertrial intervals and between-subject designs. The corresponding outcome (food reinforcement or nonreinforcement) was delivered after satisfaction of a fixed-ratio 10 (Experiment 1) or a fixed-ratio 1 (Experiment 2). Experiment 3 reports a reversed PREE in a within-subject design with a fixed-ratio 10 requirement. Extinction occurred faster for the response paired with 50% partial reinforcement than for the response paired with continuous reinforcement. A third response paired with a small reinforcer (1 pellet/trial) in 100% of the trials extinguished faster than a response paired with a large reinforcer (15 pellets/trial). These results are discussed in the context of spaced-trial instrumental performance (key pecking and running), in pigeons. Ó 2002 Elsevier Science (USA). All rights reserved. * Corresponding author. Fax: 817-257-7681. E-mail address: m.papini@tcu.edu (M.R. Papini). 0023-9690/02/$ - see front matter Ó 2002 Elsevier Science (USA). All rights reserved. PII: S 0 0 2 3 - 9 6 9 0 ( 0 2 ) 0 0 0 0 6 - 1 486 M.R. Papini et al. / Learning and Motivation 33 (2002) 485–509 Responses reinforced only in a random subset of acquisition trials typically display greater resistance in extinction, compared to continuously reinforced responses. Similarly, Pavlovian procedures in which a conditioned stimulus is partially reinforced may result in a greater behavioral level during extinction, compared to a stimulus reinforced in every trial. This phenomenon, called the partial reinforcement extinction effect (PREE), has been generally studied in between-subject designs and under relatively massed training conditions (e.g., intertrial intervals, ITI, ranging from a few seconds to a few minutes). Under such conditions, the PREE is a ubiquitous learning phenomenon both across species and training procedures (for a review, see Amsel, 1992). A different picture emerges when these procedural features are altered to within-subject comparisons between partially and continuously reinforced responses or stimuli, or to spaced training conditions with ITIs in the order of hours. Within-subject experiments have demonstrated a range of effects. For example, whereas Rescorla (1999) found a within-subject PREE (wPREE) in autoshaping with pigeons, others reported nonsignificant effects in autoshaping with quail (Crawford, Steirn, & Pavlik, 1985), salivary conditioning with dogs (Sadler, 1968), runway performance in rats (Amsel, Rashotte, & MacKinnon, 1966), and goal tracking with rats (Pearce, Redhead, & Aydin, 1997). Still others reported evidence of a reversed PREE, that is, greater persistence after training with continuous reinforcement, rather than partial reinforcement. Reversed PREEs were found in experiments with rats involving running and lever pressing responses (Pavlik & Carlton, 1965; Pavlik, Carlton, & Hughes, 1965), and in experiments with human subjects (Svartdal, 2000). A within-subject experiment reported by Mellgren and Seybert (1973) suggests one factor that might account for this diversity of results, namely, the sequence of reinforced (R) and nonreinforced (N) trials. Each rat received training in two different runways, one paired with 50% partial reinforcement and the other with continuous reinforcement. Mellgren and Seybert found the following pattern of results: A PREE occurred when N trials were followed by R trials in the partial runway; a reversed PREE emerged when N trials were followed by R trials in the continuous runway; and nondifferential extinction followed acquisition training in which the two NR transitions were equally often. The between-subject PREE (bPREE) is also affected by the spacing of the trials. Experiments with nonmammalian species that demonstrate a conventional bPREE under massed conditions of training, have consistently found a reversed bPREE under spaced training conditions. For example, toads (Bufo arenarum) trained in a runway situation to collect water in the goal box show greater persistence in extinction after partial reinforcement when trials are separated by 15-s ITIs, but not when trials are separated by 5-min ITIs (Muzio, Segura, & Papini, 1992). Moreover, lengthening the ITI to 24 h produces clear evidence of a reversed bPREE (Muzio et al., 1992, 1994). Similar reversed bPREEs have been reported in fish (Boitano & Foskett, M.R. Papini et al. / Learning and Motivation 33 (2002) 485–509 487 1968; Gonzalez & Bitterman, 1967; Schutz & Bitterman, 1969), turtles (Gonzalez & Bitterman, 1962; Pert & Bitterman, 1970), and iguanas (Graf, 1972), all trained under spaced-trial conditions. In the experiments reported in the present article, pigeons received training under conditions of spaced practice, with trials separated by ITIs in the order of hours, during which the animal remained in its home cage. In addition, both between-subject (Experiments 1 and 2) and within-subject designs (Experiment 3) were used. Our interest in the PREE is primarily comparative; the spaced-trial bPREE has been reported in several experiments with mammals (e.g., Weinstock, 1954), but not in experiments involving fish, amphibians, or reptiles (see references above). Our presumption is that these species differences in spaced-trial learning phenomena are markers of a broad phylogenetic divergence in brain mechanisms of learning among the vertebrates (Papini, 2002). The goal of the present experiments is to extend the study of the spaced-trial PREE to an avian species. Pigeons are ideally suited for our purpose because there are well-developed training techniques that can be readily applied to the present requirements. We have developed a key-pecking, one-trial-per-session situation that has been used to study the effects of reward magnitude on learned performance (Papini, 1997; Papini & Thomas, 1997); here we extend that procedure to a partial reinforcement situation, for which there is limited information. Only one demonstration of the spaced-trial bPREE in pigeons has been published and it involves training in a runway situation, 24 h ITI, and food reinforcement as the main training parameters (Roberts, Bullock, & Bitterman, 1963). Thus the present experiments extend the study reported by Roberts et al. to a key-pecking training situation and to a within-subject design. We report the results of each experiment leaving the discussion of their theoretical relevance for the General discussion section. Experiment 1 The main goal of this experiment was to determine the effect of 50% partial and 100% continuous reinforcement training on the extinction of key-pecking instrumental performance under spaced conditions of training. The basic training procedure involves a single trial per day in which a pigeon is exposed to a key light and required to peck 10 times to obtain an outcome. This outcome was either fifteen 45-mg pigeon pellets in reinforced trials, or nothing in nonreinforced trials. An analogous experiment with rats as subjects demonstrated a significant bPREE in lever-pressing instrumental performance (McNaughton, 1984). Furthermore, that experiment showed that treatment with the anxiolytic chlordiazepoxide eliminated the lever-pressing bPREE, just as it also eliminated a bPREE based on runway performance, suggesting common underlying mechanisms across these two training situations. 488 M.R. Papini et al. / Learning and Motivation 33 (2002) 485–509 The present experiment sought to determine the presence of a second effect that has been reported in experiments involving partial reinforcement under more massed conditions of training. Responses acquired under partial reinforcement training are usually performed at a higher level than responses acquired under continuous reinforcement during the acquisition phase (Goodrich, 1959). This phenomenon, labeled the partial reinforcement acquisition effect (PRAE), has been described in rats and under relatively massed conditions of training, but its occurrence in other species and under spaced training conditions has not yet been reported. Method Subjects Twelve pigeons purchased from Ruthardt Pet and Feed, Fort Worth, all sexually mature, served as subjects. Pigeons were deprived of food to an 80– 85% of their ad libitum weights. Post-session meals (at least 20 min after the session) were adjusted to maintain constant target weights. Water was continuously available in the cage. The vivarium was continuously illuminated. These pigeons had extensive prior experience in analogous experiments manipulating reinforcer magnitude (Papini, 1997, Experiments 1 and 3; Papini & Thomas, 1997). Pigeons that had received continuous reinforcement training with large rewards were assigned to the same condition in the present experiment (Group C). Pigeons that had received training with continuous reinforcement and small rewards were assigned to partial reinforcement in the present experiment (Group P). The decision to assign pigeons in agreement with their prior training history, rather than using random assignment or matching for training history, was based on two considerations. First, random or matched assignment would result in a mixture of reinforcement histories in each group that could potentially increase within-subject variance and consequently reduce power. Second, since the pigeons assigned to partial reinforcement had exhibited faster extinction in the previous experiments than those assigned to continuous reinforcement, this procedure is conservative for a demonstration of the bPREE. A demonstration of the bPREE would require that pigeons that exhibited fast extinction in previous experiments exhibit slow extinction in the present experiment, in both cases relative to the same comparison group (Group C). Apparatus Three standard Skinner boxes, each enclosed in a sound-attenuating chamber, were used for training. Each chamber was equipped with a fan to provide for air circulation and masking background noise. Each Skinner box measured 32:2 29:9 32:2 cm (width length height). The front wall, back wall, and ceiling were made of Plexiglas, whereas the two lateral walls were made of aluminum. One of the lateral walls contained M.R. Papini et al. / Learning and Motivation 33 (2002) 485–509 489 the following elements: A lamp (General Electric 1820), providing diffuse illumination from the upper left corner, and the response key (1.8 cm in diameter, 18.5 cm above the floor, and in the center of the wall). There was also a feeder cup located 3 cm above the floor and in the center of the wall; this cup was made of opaque Plexiglas and measured 4:5 5:5 4 cm: Noyes precision pellets (pigeon formula, 45 mg), were automatically delivered into this cup. Visual stimuli could be presented on the response key by illuminating it from behind. A computer located in an adjacent room controlled all the stimuli and recorded the key-pecking responses. Procedure Experimental sessions were conducted between 08:00 and 12:00 h. Pigeons were extensively pretrained. First, all pigeons received 2 daily sessions of exposure to the conditioning box, each lasting 20 min; no stimuli were presented during these sessions. Second, birds were exposed to a mixed Pavlovian-instrumental schedule (20 trials/session). In the first session of this series, a small amount of mixed grain (about 2 g) was freely available in the food cup; no grain was provided thereafter. A single key-peck at a white key light or 6 s, whichever occurred first, resulted in the termination of the light and the immediate delivery of a single pellet. Each pigeon was required to respond on at least 80% of the trials on two of three consecutive days before it was shifted to the next phase. Third, after achieving this criterion, purely instrumental pretraining was initiated in which reinforcement occurred only if the pigeon responded during the 15-s long white key light. During these sessions, the fixed-ratio requirement was gradually increased from 1 to 10 responses, in steps of one key-peck. Before each transition to the next value, pigeons had to satisfy the same criterion described above. A transition from fixed-ratio 10 to a single trial per day (acquisition phase, see below) required 80% or better response level in three of five consecutive sessions. This gradual procedure was adopted to minimize behavioral disruptions and to insure the occurrence of key pecking after the shift from 20 trials/session to 1 trial/session. This pretraining procedure was used in previous experiments (Papini, 1997; Papini & Thomas, 1997). One trial/day was administered to each animal during acquisition. Pigeons assigned to Group C (n ¼ 6) received 15 pellets at the end of each trial, delivered at a rate of 1 pellet every 0.2 s, provided they had emitted 10 responses. For the pigeons assigned to Group P (n ¼ 6), half of the daily trials ended in reinforcement (15 pellets delivered as described above), whereas the rest ended in nonreinforcement. The sequence of R and N trials was based on Gellermann (1933): RRNN RNRN NRRR NNRN NRNR RNRN RRNN RRNR NNRR NNRR NRNN NRRN RRNN. Thus the labels ‘‘continuous’’ and ‘‘partial’’ apply to the outcome of emitting 10 responses, and not to the outcome of each individual response (because 10 key pecks were required on each trial), or to the stimulus-reinforcer 490 M.R. Papini et al. / Learning and Motivation 33 (2002) 485–509 contingency (because failure to complete the fixed-ratio 10 ended the trial in nonreinforcement). In this acquisition phase, the stimulus was changed from white (pretraining) to a black plus sign on a white background. A trial started with onset of the house light, followed by an initial interval averaging 60 s (range: 30–90 s). At the end of this interval, the plus stimulus was projected on the key; animals were allowed a maximum of 60 s to initiate responding and an additional 60 s to complete the fixed-ratio 10. A computer located in an adjacent room recorded the latency (0.01 s units), defined as the time from stimulus onset to the tenth response. Failure to complete the fixed-ratio 10 within 60 s ended the trial in nonreinforcement. A maximum latency of 120 s was assigned on incomplete trials. The session ended after a variable interval similar to the initial one. At the end of this interval, the house light was turned off and the pigeon was returned to its home cage. There were 52 acquisition trials followed by 88 extinction trials; pigeons receive one trial per day. Extinction trials were identical to the nonreinforced trials of Group P. Latency scores obtained in each trial and measured in 0.01 s units were transformed to the natural logarithm (ln), to improve normality and allow for the use of parametric statistics. A standard, mixed-model analysis of variance was used to test the effects of reinforcement schedule on the latencies. The alpha value was set to 0.05 in all the analyses reported in this paper. Nonparametric tests were also used when extreme numbers skewed data distributions. Results Given the criteria used to increase the fixed-ratio value, pretraining required a minimum of 24 sessions. Eleven pigeons finished pretraining in 24 sessions, while the remaining pigeon, from Group P, required a total of 28 sessions. Although pretraining was extensive, it was not differentially so across groups. Moreover, none of the pigeons exhibited behavioral disruption during pretraining sessions. There was no evidence of the PRAE in the present experiment. Key-pecking performance during acquisition was similar across groups. The change in stimulus from pretraining (white key-light) to acquisition (plus sign) resulted in a small degree of disruption in key pecking during the early acquisition trials. None of the effects of a Group Trial analysis of variance was significant, F s < 2:97. An analysis of the last 8 acquisition trials also indicated a nonsignificant group effect, F ð1; 11Þ ¼ 1:11; thus, asymptotic performance was very similar across groups. As shown in Fig. 1, initial extinction performance was extremely stable and similar in both groups. An analysis of the initial 8 extinction trials indicated that groups were statistically indistinguishable, F < 1. Eventually, the pigeons in Group C extinguished faster than those in Group P. A Group Trial analysis (computed on a trial-by-trial basis), detected this M.R. Papini et al. / Learning and Motivation 33 (2002) 485–509 491 Fig. 1. Extinction of key-pecking performance in groups of pigeons trained with either continuous (C) or 50% partial (P) reinforcement, at a rate of a single trial per day. A fixed-ratio 10 requirement was imposed in each trial. Data from Experiment 1. bPREE in terms of a significant interaction, F ð87; 870Þ ¼ 1:33. There was also a significant extinction effect, F ð87; 870Þ ¼ 38:65, but the difference between groups was not significant, F ð1; 10Þ ¼ 3:51. The interpretation of this bPREE is not affected by initial group differences in performance level, as shown by nonsignificant group effects for the final acquisition trials and initial extinction trials. The instrumental procedure used in this experiment implied that an animal nominally subjected to 100% continuous or 50% partial reinforcement could receive a lower frequency of reinforcement (in the sense of the stimulus-reinforcer contingency), if it failed to complete the fixed-ratio 10 on a substantial number of trials. Disruption was assessed in terms of the number of trials on which pigeons from each group failed to complete the fixed-ratio 10 during the time allowed, during both acquisition and extinction sessions. The degree of disruption in acquisition was relatively low, averaging 1.8 trials in Group C and 0.2 trials in Group P; with a total of 52 acquisition trials, pigeons completed more than 96% of all acquisition trials. In extinction, pigeons show disruption in an average of 49.7 and 26.3 trials in Groups C and P, respectively (out of a total of 88 trials). Because 8 of the 12 animals showed no disruption in acquisition, all disruption data were analyzed using nonparametric tests. Mann–Whitney tests indicated that disruption during acquisition was not significantly different across groups, U ð6; 6Þ ¼ 11, and there was a tendency toward a greater disruption in Group C than in Group P during extinction, significant only at a one-tail test level, U ð6; 6Þ ¼ 5:5. 492 M.R. Papini et al. / Learning and Motivation 33 (2002) 485–509 Experiment 2 There are two potential problems with the results of Experiment 1. First, since 10 key pecks were required to deliver the reinforcer, it may be argued that the pigeons from both groups were under a partial reinforcement regime. The main difference may have been that whereas the pigeons in Group C received a reinforcer for every 10 responses, those in Group P had to emit 20 responses, on average, to be reinforced. To clarify this issue, two new groups of animals were run under the same conditions as those in the previous experiment, except that a single peck at the key was required for reinforcement (i.e., a fixed-ratio 1 requirement). Second, in Experiment 1 and for reasons that were explained previously (see Experiment 1, Subjects), group assignment was consistent with the previous experience of the pigeons (e.g., prior large reward was assigned to large reward; prior small reward was assigned to partial reward). The pigeons used in the present experiment also had previous experience (see Stout, Muzio, Boughner, & Papini, 2002, Experiments 4 and 5), in a complex training situation. In those experiments, responses to two stimuli (green light and plus sign), were paired with reinforcement and nonreinforcement in either a discrimination (i.e., Aþ/B), or a pseudodiscrimination (i.e., A/B) condition, across groups. These trials were followed, after a brief interval, by a white key light (to be used in the present experiment), always reinforced on a variable-interval 20-s schedule (i.e., Cþ). Two precautions were taken with these subjects. First, pigeons were matched for prior experience (discrimination vs. pseudodiscrimination training) within each of the conditions used in the present experiment (Groups C and P). Second, the white stimulus, which had been subject to the same treatment in both groups of the previous experiments, served as the discriminative stimulus in the present experiment. Method Subjects and apparatus Twelve pigeons were utilized in this experiment, all sexually mature, obtained from the same source, and maintained under the same conditions as in Experiment 1. The same Skinner boxes described in Experiment 1 were also used in the present experiment. Procedure Pairs of pigeons matched for prior experience were randomly assigned to the two conditions of the present experiment: Groups C and P ðn ¼ 6Þ. Thus, three animals in each of the present groups had previously received discrimination training and the rest had received pseudodiscrimination training. Each pigeon received one trial per day, between 08:00 and 12:00 h. In each trial, pigeons were placed in the conditioning box, the house M.R. Papini et al. / Learning and Motivation 33 (2002) 485–509 493 light was turned on, and after an average period of 60 s (range: 30–90 s), the white light was presented on the response key. A single peck at the white key turned the white light off and administered 15 pellets (in reinforced trials). Pellets were administered at a rate of 1 every 0.2 s. This reinforcement was followed by an interval averaging 60 s (range: 30–90 s), at the end of which the house light was turned off and the pigeon was returned to its cage. During the entire experiment, pigeons receive a single trial per day. In trials 1–4, all pigeons were immediately reinforced for responding within 60 s of key-light onset. Failure to respond resulted in the automatic delivery of the reinforcer after 60 s. Thus, these sessions involved a mixture of Pavlovian and instrumental contingencies designed to recover key-pecking behavior. Starting in session 5, pigeons received the appropriate training under purely instrumental conditions. Pigeons in Group C received a total of 56 trials in which a single key-peck response resulted in the delivery of the reinforcer. Pigeons in Group P received 56 trials in which R and N trials were intermixed according to the following schedule based on Gellermann (1933): RNRN RRNN NRRN RNNR RRNN RNRN NRRN NRRN RNNR NNRR RNNR RRNR NNRN NRRR. Thus, there were a total of 60 acquisition trials. Extinction started on trial 61 and continued for 48 trials. In nonreinforced trials (e.g., in N trials for Group P and extinction trials for both groups), a peck at the white key or 60 s, whichever occurred first, turned off the white light and ended the trial. The primary dependent measure was the latency to respond, the time from key-light onset to the detection of a peck (measured in 0.01 s units). These scores were treated as described in Experiment 1. Results One pigeon died in the course of acquisition training and two other pigeons failed to reacquire the response. Although it would have been possible to retrain the two pigeons that failed to respond, it was decided to discard them from the experiment to keep previous experience as constant as possible across groups. As a result, the 9 pigeons that remained in this experiment, 4 in Group C and 5 in Group P, all showed recovery of the key-pecking response when transferred to a one-trial-per-day regime without any explicit pretraining. The key-pecking performance of both groups during acquisition was similar, thus yielding no evidence of the PRAE. A Group Trial analysis indicated a significant acquisition effect, F ð59; 413Þ ¼ 1:96; however, the group effect and the interaction effect were both nonsignificant F s < 1:33. Furthermore, an analysis of the final 8 trials of acquisition indicated no difference among groups and no group trial interaction, F s < 1:16, although the latencies were still decreasing significantly, F ð7; 49Þ ¼ 3:34. The results of the extinction phase are presented in Fig. 2 for each group and as a function of 4-trial blocks. The latency scores were lower than in the 494 M.R. Papini et al. / Learning and Motivation 33 (2002) 485–509 Fig. 2. Extinction of key-pecking performance in groups of pigeons trained with either continuous (C) or 50% partial (P) reinforcement, at a rate of a single trial per day. A fixed-ratio 1 requirement was imposed in each trial. Data from Experiment 2. previous experiment simply because it takes less time to emit a single response than the time it takes to complete 10 responses. Two features of these results are worth emphasizing. First and as in the previous experiment, the initial extinction performance was relatively stable, except for some initial fluctuation in Group C. For approximately 20 trials, the latencies in both groups remained relatively low. An analysis of the initial 8 trials indicated nonsignificant group differences, F < 1. Second and also as in the previous experiment, there was clear evidence of a bPREE. Indeed, there was little change in performance during the 48 extinction trials in Group P. A trial-by-trial analysis of extinction data detected the bPREE in terms of a group trial interaction, F ð47; 329Þ ¼ 2:61. Also significant was the extinction effect, F ð47; 329Þ ¼ 4:46, but not the difference between the groups, F ð1; 7Þ ¼ 2:82. Because terminal acquisition performance and early extinction performance were not different across groups, this bPREE is not confounded with initial differences in latency. Only three of the nine pigeons failed to respond in at least one of the acquisition trials, one in Group C and two in Group P. Although the average number of disrupted trials across individual pigeons for Groups C and P (0.75 and 4.40 trials, respectively) seem different, the difference was not statistically reliable, U ð4; 5Þ ¼ 7:5. This group tendency reversed during extinction, with Group C exhibiting greater average disruption (7.50 trials), M.R. Papini et al. / Learning and Motivation 33 (2002) 485–509 495 than Group P (0.25 trials); however, this difference also failed to achieve a significant level, U ð4; 5Þ ¼ 3. Thus the bPREE found in the present experiment was not related to differential disruption in performance across groups. Experiment 3 An experiment by Mellgren and Seybert (1973) provides information on the effects of partial and continuous reinforcement on extinction in a withinsubject design and spaced conditions of training (i.e., one trial per day). Rats were trained in two different runways, one associated with partial reinforcement and the other with continuous reinforcement. Extinction was similar in the partial and continuous runways when the number of NR trial transitions was matched across runways. However, in most within-subject experiments, ITI values ranged from zero, as in free-operant situations (Pavlik & Carlton, 1965), up to a few minutes (Pearce et al., 1997). Additionally, in all of these experiments the animals remained in the conditioning chamber during the ITI. Under relatively massed training conditions, the outcome of one trial may leave a sensory or memory trace capable of remaining active until the next trial (Sheffield, 1949). As a result of such carry-over effect, responding in any given trial may fall under the control of both current external stimuli as well as carry-over stimuli (whether sensory or mnemonic). Additionally, the massing of trials may promote a comparison among stimuli paired with different frequencies of reinforcement. Such comparisons may induce contrast effects, affecting response rates in ways that may be difficult to anticipate (Svartdal, 2000). A unique advantage of spaced-trial procedures is that they reduced or eliminated these problems, promoting a relatively clean control of behavior by the associative value acquired by the discriminative stimuli. Furthermore, Papini (1997) reported evidence of a reverse magnitude of reinforcement extinction effect (MREE) in both between- and within-subject experiments with pigeons and under spaced-trial conditions. The MREE is defined as greater persistence in extinction after acquisition with a small, rather than large, reinforcer magnitude (Hulse, 1958; Wagner, 1961). Mellgren and Dyck (1974) reported a within-subject experiment with rats in which two discriminable runways were paired with large and small rewards under massed-training conditions (i.e., 15-s ITIs). Rats that had received an equal number of transitions from large to small rewards and from small to large rewards exhibited the MREE. However, there seem to be no reports of MREE studies based on spaced-training conditions and within-subject designs in rats or any species. A tentative summary of available evidence indicates that partial and small reinforcers generate relatively greater persistence in extinction under spaced-trial conditions in rats, but their effects are dissociated in spaced-trial extinction in pigeons. In fact, the spaced-trial 496 M.R. Papini et al. / Learning and Motivation 33 (2002) 485–509 PREE and MREE co-vary in any given species in which they have been studied (see Papini, 1997; Papini, Muzio, & Segura, 1995), except for pigeons. In pigeons, the spaced-trial bPREE has now been demonstrated in runway (Roberts et al., 1963; Thomas, 2001), and Skinner box situations (Experiments 1 and 2), whereas the spaced-trial reversed MREE has been found in both runway (Thomas, 2001), and Skinner box situations (Papini, 1997; Papini & Thomas, 1997). The present experiment was designed to provide further information on the dissociation of spaced-trial PREE and MREE in pigeons by testing both effects in a within-subject design. Each pigeon received exposure to three discriminative stimuli, one continuously reinforced with a large reward, a second paired with 50% reinforcement of the same large reward, and a third continuously reinforced with a small reward. Eventually, all three stimuli were shifted to extinction. Method Subjects and apparatus Nine adult pigeons, all experimentally naive, served as subjects. They were obtained and maintained as described in Experiment 1. The same Skinner boxes described above were also used in the present experiment. Procedure Experimental sessions were conducted between 08:00 and 12:00 h. Pretraining was exactly as described in Experiment 1. Pigeons were gradually shaped to peck at the white key light for a single pellet of food, and daily sessions contained 20 trials each. The response requirement was gradually increased up to a fixed-ratio 10 according to the same behavioral criteria described in Experiment 1. Subsequently, pigeons were shifted to the acquisition phase. During acquisition, animals received 1 trial per session and 3 sessions per day. Each trial was exactly as described in Experiment 1. At the end of each trial, the animal was returned to its cage where it remained during the ITI. Within any given day, the ITI was 1–1.5 h long; between days, the ITI was between 18- and 21-h long. Three different stimuli were presented to each pigeon, although only one in any particular trial. The stimuli were a white plus sign on a black background, a red key, and a green key. For any given animal, responding 10 times to each stimulus produced one of the following outcomes: delivery of 15 pellets in every trial (the continuous, large-reward stimulus, or CL), delivery of 15 pellets or nothing on a random 50% basis (the partial, large-reward stimulus, or PL), or delivery of 1 pellet in every trial (the continuous, small-reward stimulus, or CS). The stimulus-outcome relationship was counterbalanced across animals, such that each stimulus was paired to each of the three outcomes in an equal number of animals. M.R. Papini et al. / Learning and Motivation 33 (2002) 485–509 497 Acquisition trials were scheduled in 3-day blocks. In each 3-day block, a given animal received 3 trials of each type (CL, PL, and CS), distributed randomly except that no more than 2 trials of the same type were administered in any given day. Three sequences were administered, each to 3 pigeons: PLP SLL SSP, LSL PSS PPL, and SPS LPP LLS (where P is the partial stimulus, which was R and N on a 50% schedule; L is the CL stimulus; and S is the CS stimulus). This procedure ensured that each stimulus was presented the same number of times in each possible position within a day for the group as a whole. There were 48 acquisition trials of each type, followed by 48 extinction trials of each type. A total of 288 trials were administered in the entire experiment. Extinction trials were exactly the same as acquisition trials, except that no food was delivered. No less than 1 h after the end of the third trial, in any given day, pigeons received sufficient supplementary food to maintain their body weights at 80–85% of their free-food weight. The latency to complete the fixed-ratio 10 requirement (time from the onset of the stimulus until the emission of the 10th key-peck response, in 0.01 s units) was the primary dependent measure. Scores were treated as described in Experiment 1. For consistency with the previous experiments, disruption data were analyzed with nonparametric techniques. Results An average of 33.2 (range 29–35), sessions of pretraining were needed to shape the fixed-ratio 10 requirement. None of the pigeons exhibited any behavioral disruption during pretraining. Acquisition performance was very similar across trial types. A Trial Type (CL, PL, CS), by Trial analysis of variance (both as repeated-measure factors), indicated only a significant acquisition effect, F ð47; 376Þ ¼ 3:27. The difference across trial types and the interaction effect were nonsignificant, F s < 1:04. The same results were obtained in a pairwise comparison between CL and PL trials (partial reinforcement), and between CL and CS trials (magnitude of reinforcement). The acquisition effect was significant in both of these analyses, F sð1; 8Þ > 2:96, but none of them detected a significant trial-type effect, F s < 1, or interaction effect, F sð47; 376Þ < 1:15. Performance during the asymptotic period was also very similar across trial types. An analysis of the last 8 acquisition trials indicated nonsignificant differences between CL and PL trials, F < 1, and between CL and CS trials, F < 1. The extinction performance of these pigeons in the three types of trials is shown in Fig. 3. Behavioral changes start comparatively earlier than in the previous experiments, suggesting a substantial generalization of extinctionÕs decremental effects across stimuli. Nonetheless, an analysis of the initial 8 trials of extinction shows nonsignificant differences in behavior between CL and PL trials, F ð1; 8Þ ¼ 3:47, and between CL and CS trials, F < 1. Thus the performance across trial types in both late acquisition and early 498 M.R. Papini et al. / Learning and Motivation 33 (2002) 485–509 Fig. 3. Extinction of key-pecking performance in three different discriminative stimuli associated with continuous-large reward (CL), 50% partial-large reward (PL), and continuous-small reward (CS). Training was administered at a rate of 3 trials per day with a minimum ITI of one hour. Animals remained in their cages during the ITI. A fixed-ratio 10 requirement was imposed in each trial. Data from Experiment 3. extinction was statistically indistinguishable, a fact that simplifies the interpretation of extinction data. Overall, Fig. 3 suggests that the extinction performance in CL trials was consistently below that of either PL (reversed wPREE), or CS trials (reversed wMREE). An overall analysis indicated significant differences across trial types, F ð2; 16Þ ¼ 5:76, and trials, F ð47; 376Þ ¼ 27:54, but a nonsignificant interaction, F ð47; 376Þ ¼ 1:11. Two pairwise comparisons are relevant in this experiment. First, a comparison of CL and PL trials indicated significantly lower extinction latencies in the continuous, rather than partial, reinforcement stimulus, F ð1; 8Þ ¼ 6:88, thus demonstrating a reversed wPREE. Second, a similar comparison between the CL and CS stimuli indicated significantly lower extinction latencies after training with large, rather than small, reinforcement, F ð1; 8Þ ¼ 10:79, revealing a reversed wMREE. In both of these analyses there were significant extinction effects, F sð47; 376Þ > 20:45, but nonsignificant interaction effects, F sð37; 376Þ < 1:26. None of the nine pigeons exhibited disruption of key-pecking performance in any of the 144 acquisition trials. During extinction, pigeons completed the fixed-ratio 10 requirement in an average of 30.7 CL, 22.9 PL, and 22.8 CS trials. Wilcoxon tests (two-tailed) demonstrated that pigeons M.R. Papini et al. / Learning and Motivation 33 (2002) 485–509 499 completed a significantly larger number of CL trials than both PL trials, T ð9Þ ¼ 0, and CS trials, T ð9Þ ¼ 1. Thus these results demonstrate the presence of reversed wPREE and wMREE with disruption as the dependent measure. General discussion Two main findings stem from the present experiments. First, Experiments 1 and 2 provide an original demonstration of the spaced-trial bPREE in a key-pecking procedure with pigeons. Together with an analogous runway experiment reported by Roberts et al. (1963) and with additional experiments from our lab (Thomas, 2001), they constitute the only evidence available of the spaced-trial bPREE in any nonmammalian species; this result is thus particularly important from the comparative point of view. Experiment 3 provides the second main contribution of this paper. As far as we know, this is the first report to study the effects of both reinforcer schedule and reinforcer magnitude in a within-subject, spaced-trial design. As previously mentioned, massed-training procedures allow for stimulus and memory carry-over effects from prior trials to control the instrumental response (Sheffield, 1949). In addition, within-subject designs involving relatively massed conditions of training encourage comparison across training conditions that may influence behavior (Svartdal, 2000). These potential carry-over and comparison effects are minimized or eliminated by the use of widely spaced training conditions and by the removal of the animals from the conditioning context during the ITI. As a result, spaced-trial procedures permit a cleaner assessment of the manner in which the associative strength of the discriminative stimulus controls the instrumental response. The pattern of results reported in these experiments (i.e., bPREE and reversed wPREE) is therefore not attributable to these factors. Different extinction performance across groups or conditions in an experiment may sometimes be attributable to the effects of the target manipulations on acquisition. Different transformations have been suggested to deal with the problem generated by group differences in early extinction trials (e.g., Anderson, 1963). However, the bPREE reported in Experiments 1 and 2, and the reverse wPREE reported in Experiment 3 occurred in the absence of an effect of partial reinforcement training on acquisition in general, and on late acquisition or early extinction in particular. The bPREEs obtained in Experiments 1 and 2 could be attributed to the prior experience of the pigeons or, perhaps, the reverse wPREE reported in Experiment 3 could be the result of using experimentally naive animals. A consideration of the assignment procedures, which were different in each of these experiments, suggests that prior experience is unlikely to be the source of these results. Whereas the assignment of pigeons in Experiment 1 was 500 M.R. Papini et al. / Learning and Motivation 33 (2002) 485–509 congruent with prior training, the assignment in Experiment 2 was orthogonal to previous experience. Furthermore, in Experiment 1, for which assignment was not orthogonal, the results of the previous and present experiments were opposite. Thus, the pigeons showing the slowest extinction in Experiment 1 (Group P) were those showing the fastest extinction in their previous experiments (small reward group in Papini, 1997, Experiments 1 and 3, and in Papini & Thomas, 1997). Also relevant is the fact that the previous experience of the pigeons used in Experiments 1 (mostly one-trial-perday experiments; Papini, 1997; Papini & Thomas, 1997), and 2 (massedtraining procedures; Stout et al., 2002) was considerably different. Additional information suggesting that the bPREE is not an artifact of prior training comes from studies involving experimentally naive pigeons. Using a runway situation, Thomas (2001) found that reward inconsistency increased persistence in extinction of the running response both after partial reinforcement training (thus replicating the results reported by Roberts et al., 1963), and after a random mixture of large and small rewards. It is thus safe to conclude that reward uncertainty in acquisition increases persistence in extinction when pigeons are trained under spaced-trial conditions and comparisons are made between groups. Although currently popular theoretical accounts may explain one of the reported effects, most theories would find it difficult to account for the complete pattern. Moreover, the picture is more complex if the effects of reward magnitude are added. The available evidence on spaced-trial effects in pigeons is summarized in Table 1. There are two basic problems to be reconciled, both involving a dissociation of results. The first dissociation involves the regular bPREE and the reversed MREE, found in both between- and within-subject designs. The second dissociation involves the Table 1 Summary of results from spaced-trial experiments with pigeons PREE (Reference) Between-subject designs Key-Pecking Conventional (Present Experiments 1–2) Running Conventional (Roberts et al., 1963) (Thomas, 2001, Experiment 1) Within-subjects designs Key-Pecking Reversed (Present Experiment 3) Running No effect (see text for details) (Mellgren & Seybert, 1973) MREE (Reference) Reversed (Papini, 1997, Experiments 1–2) (Papini & Thomas, 1997) Reversed (Thomas, 2001, Experiment 1) Reversed (Papini, 1997, Experiment 3) (Present Experiment 3) No information available M.R. Papini et al. / Learning and Motivation 33 (2002) 485–509 501 PREE, which appears in several between-subject experiments, but not in the within-subject experiment reported here, which yielded evidence of a reversed wPREE. Consider the first dissociation. A basic problem with the dissociation of spaced-trial bPREE and reversed MREE is that schedule and magnitude of reinforcement have traditionally been thought of as affecting associative strength in the same manner. One example is the simple view, first suggested by Thorndike (1911), that reinforcement strengthens a stimulus-response association, whereas nonreinforcement weakens it. This theoretical notion has been applied extensively with some success (e.g., Couvillon & Bitterman, 1985; Rescorla & Wagner, 1972; Schmajuk, 1997). According to this strengthening–weakening (S–W) view, the strength of a stimulus increases directly with the frequency and magnitude of reinforcement. Most of these models therefore predict reversed bPREEs and MREEs in any situation in which the associative strength of a stimulus can be assessed relatively directly, as in spaced-trial situations. The failure of S–W views led to the development of other theories, which, although based on different principles, still predict that training conditions producing one effect should also produce the other. For example, AmselÕs frustration theory (1992) and CapaldiÕs sequential theory (1994), both considered classic accounts of the PREE, can explain the results of Experiment 1 in general but fail to predict more specific aspects of such results. AmselÕs, 1992, account is inconsistent with the absence of a PRAE, and both accounts would demand a faster extinction rate than that observed in Group C. As shown in Fig. 1, the behavior of continuously rewarded pigeons remained undisturbed during the initial 24 trials of extinction. Moreover, both of these classic theories fail to account for the reversed wPREE obtained in Experiment 3 and for the reversed bMREE and wMREE (e.g., Papini, 1997; present Experiment 3). In terms of frustration theory, a within-subject design should lead both the partial and continuous cues (and the large- and small-reinforcer cues) to induce a common internal response of anticipatory frustration, which should result in nondifferential extinction to both cues (e.g., Amsel et al., 1966). The second dissociation (i.e., bPREE versus reversed wPREE), is also difficult to explain for most learning theories. A viable account of this dissociation was offered by Mellgren and Seybert (1973), based on CapaldiÕs sequential theory. For a within-subject experiment, sequential theory predicts nondifferential extinction rates for stimuli paired with partial-large reinforcement and continuous-large reinforcement, provided a specific assumption is met, namely, that there is an equal number of NR transitions in the partial-large stimulus and in the continuous-large stimulus. When this provision is not met, the results may lead to either a wPREE or a reversed wPREE depending on the sequential bias of the schedule used in acquisition, as mentioned previously (Mellgren & Seybert, 1973). Because the 502 M.R. Papini et al. / Learning and Motivation 33 (2002) 485–509 sequence of the three stimuli used in Experiment 3 was counterbalanced across subjects, different pigeons received a different number of transitions. The counterbalance criterion was chosen to equate the number of times each stimulus appeared in each position during a series of 9 successive trials (3 trials/day). According to this procedure, PL, CL, and CS trials each occurred three times in each of the three positions (first, second, or third trial of the day) for the group as a whole. Sporadic errors that modified the scheduled sequence for a particular animal also introduced variability across pigeons. Table 2 shows the results of an analysis of the actual sequences per stimulus computed separately for the sequences occurring within days and for those occurring across days. Data on transitions from small reward to large reward were added because sequential theory clearly applies to such cases (Capaldi, 1994). In agreement with Mellgren and SeybertÕs (1973) results, the continuous-large stimulus exhibited both the highest number of critical transitions (within and between days) and the highest persistence in extinction of the three stimuli used in Experiment 3. Moreover, strong correlations were also evident across pigeons in terms of these two measures, but their sign was opposite for within- and between-day sequences. For the trials occurring within a single day and separated by an ITI of about 1–1.5 h the correlation was positive, as predicted by sequential theory, namely, the larger the number of transitions, the greater the level of persistence in extinction. However, for sequences occurring across days and separated by an ITI of 18–21 h, the correlation was significant but negative. The results for the partial-large and continuous-small stimuli also contradicted sequential theoryÕs predictions. Despite individual variation in both measures, none of the correlations approached significance and, in two cases, their sign was negative. Of course, the sequential hypothesis also predicts the MREE based on the assumption that, in extinction, there would be greater Table 2 Sequential analysis of Experiment 3 Sequences within days Probability of response PearsonÕs coefficient Sequences between days Probability of response PearsonÕs coefficient Partial large NL þ SL Continuous large NL þ SL Continuous small NS 7.3 (2–12) 0.48 (0.21–0.79) )0.41 16.6 (13–19) 0.64 (0.29–0.92) 0.80 4.8 (3–7) 0.47 (0.13–0.85) )0.42 4.3 (0–7) 0.48 (0.21–0.79) 0.60 7.9 (4–13) 0.64 (0.29–0.92) )0.88 2.9 (0–6) 0.47 (0.13–0.85) 0.64 Note. N: nonreinforcement in partial-large trials. L: large reinforcer. S: small reinforcer. The sequences of reinforcement and nonreinforcement correspond to the acquisition phase. The probability of response was computed in terms of the number of extinction trials in which the pigeon completed the fixed-ratio 10 requirement for each stimulus divided by 48 (the total number of extinction trials with each stimulus). Ranges are given in parentheses. * p < 0:01, two-tailed test. M.R. Papini et al. / Learning and Motivation 33 (2002) 485–509 503 generalization of memory control from a large reward to nonreward, than from a small reward to nonreward (Capaldi, 1994). This prediction is not supported by spaced-trial experiments that have consistently shown reversed bMREE and wMREE (Papini, 1997; Papini & Thomas, 1997). Furthermore, and based on the same memory generalization-decrement mechanism, sequential theory also predicts that a transition from a large to a small reinforcer magnitude should produce a successive negative contrast effect, which failed to develop also under spaced-trial conditions (Papini, 1997). More recent theoretical arguments also find it difficult to account for the emerging pattern of results from spaced-trial experiments with pigeons. For example, Pearce et al. (1997), elaborating on CapaldiÕs (1994) sequential theory, suggested that the conditions of reinforcement and nonreinforcement induce an internal state that accompanies the presentation of cues thus providing an internal context. In a between-subject situation, the internal context changes more abruptly in extinction for animals exposed to continuous reinforcement, than for animals trained with partial reinforcement, thus giving rise to the conventional bPREE. However, a within-subject design implies a common internal context thus equating generalization from acquisition to extinction for both the continuous and the partial cues. Under these conditions, this account predicts that performance in extinction would reflect the associative strength of each cue. Given that the partial cue has been frequently nonreinforced, it is expected to have less strength than the continuous cue hence leading to a reversed wPREE. On the assumption, mentioned previously, that reinforcement frequency and magnitude have similar effects, then a cue paired with a small reward should acquire less strength than one paired with a large reward. Pearce et al.Õs (1997) account works well with the results of the present experiments, but it fails to predict the reversed bMREE reported in previous articles (Papini, 1997; Papini & Thomas, 1997). The bPREE and reversed bMREE were also obtained within the same experiment, measuring running performance (Thomas, 2001). The results of spaced-trial experiments with pigeons are also inconsistent with other contemporary proposals. First, Daly and DalyÕs (1982) model, a set of linear equations combining Rescorla-WagnerÕs formal account of conditioning and AmselÕs frustration theory, accounts for the bPREE observed in Experiment 1, but predicts conventional wPREE and wMREE for Experiment 3, rather than the reversed effects that were actually obtained. Second, Eisenberger (1992) suggested that partial reinforcement induces a type of generalized persistence, called learned industriousness, which transfers from the original training context to new situations. This account leads to the expectation that extinction should be equivalent in the within-subject situation; presumably, the persistence induced by the partially reinforced cue should transfer to all other cues extinguished in the same or in a different situation. Precisely the opposite was found in the present Experiment 3. Third, Rescorla (1999), elaborating on AmselÕs frustration theory, argued 504 M.R. Papini et al. / Learning and Motivation 33 (2002) 485–509 that performance in within-subject conditions would be under the control of a configuration of the external cue plus an internal condition such as anticipatory frustration (see also Pearce et al., 1997). A configural control of this type is at variance with the reversed wPREE and wMREE reported in Experiment 3. An explanation of the apparent dissociation between PREE and MREE found in pigeons would seem to require a mechanism that applies to reinforcer schedule but not to reinforcer magnitude. Ideally, such a mechanism would also have to be sensitive to the between-subject versus within-subject dimension to successfully deal with the second dissociation mentioned previously (i.e., between-subject versus within-subject effects). Following Pearce et al.Õs (1997) suggestion, one could argue that when this putative mechanism is engaged, it takes control of behavior away from the basic S–W process but, when it is not engaged, the S–W mechanism operates freely. Thus, the bPREE would be produced by a mechanism engaged by reinforcement uncertainty that overlies the S–W process, whereas the reversed MREE and reversed wPREE would be produced by the simple S–W mechanism. There are at least three hypotheses that may potentially fit this scenario. The first is referred to as the response-unit hypothesis. Mowrer and Jones (1945) suggested that partial reinforcement shapes a larger behavioral unit than that shaped by continuous reinforcement. For example, an animal receiving one reinforcer per response develops a unit involving one response, whereas another receiving, on average, one reinforcer every other response develops a unit involving two responses. When shifted to extinction, the second animal will emit twice as many responses than the first, but the same number of units. There is some evidence that reinforcement can induce the formation of response patterns that exhibit a unified organization. For example, pigeons required to peck four times, in any order, on two response keys tend to develop a stereotyped sequence. When such sequences are exposed to fixed-interval or fixed-ratio schedules, the sequence as a whole exhibits properties that are typical of individual responses, such as a postreinforcement pause (Schwartz, 1982). The response-unit hypothesis predicts that the bPREE should disappear (i.e., partial and continuous groups exhibit the same extinction performance) when response units are matched across groups. Although this hypothesis was not explicitly postulated to explain the effects of reinforcer magnitude on extinction, it could be easily extended to such a case. Applied to the case of reinforcer magnitude, the response-unit hypothesis would plausibly predict that the response unit will be established faster in the large-reinforcer condition than in the small-reinforcer condition. This follows from the application of the S–W rule to the response unit. In turn, this difference would lead to the expectation that the large-reinforcer group would extinguish more slowly than the small-reinforcer group when extinction is measured in terms of individual responses. This prediction fits the available evidence for pigeons (Papini, M.R. Papini et al. / Learning and Motivation 33 (2002) 485–509 505 1997; Papini & Thomas, 1997; present Experiment 3). Furthermore, the response-unit hypothesis predicts that the reversed MREE should diminish or disappear after extended acquisition practice, a prediction that is both counterintuitive and easy to test. The failure of the response-unit hypothesis to explain the absence of dissociation in the within-subject condition, which yielded a reversed wPREE and a reversed wMREE, could be accommodated in the following manner. In a within-subject situation, response units could only develop if they can fall under stimulus control. It seem plausible that the intermixing of trial types in a within-subject situation would introduce substantial levels of proactive interference that would tend to disrupt the development of response units. Proactive interference has been demonstrated in pigeons in situations involving exposure to multiple stimuli (e.g., Roberts, 1980). The response-unit hypothesis can be accommodated to explain the entire pattern of results described in Table 1 and therefore merits a more direct testing. The second may be referred to as the timing hypothesis. Gibbon, Farrell, Locurto, Duncan, and Terrace (1980) suggested that trial responding depends on a comparison between the food expectancy induced by trial stimuli versus that induced by the context in which trials take place. They further assumed that these expectancies are inversely related to the duration of the trial stimulus and directly related to the average interreinforcement interval, respectively. Thus, since the animal receives, on average, twice as many trials per reinforcement at twice the interreinforcement interval in a 50% schedule than in a 100% schedule, the ratio of trial to interreinforcement durations is equal in both situations. The timing hypothesis predicts that the bPREE would disappear when extinction performance is measured as a function of expected reinforcements, rather than trials, a prediction which Gibbon et al. (1980) confirmed in an extensive reanalysis of published evidence from Pavlovian conditioning experiments (see also Gallistel & Gibbon, 2000; Gibbon & Balsam, 1981). This timing hypothesis was developed to a large extent on the basis of data collected in autoshaping studies with pigeons. Unlike in the present procedure (based on the delivery of discrete pellets), autoshaping studies have typically varied reinforcer magnitude in terms of hopper duration, the time during which a hopper containing grain is made accessible to the pigeon (e.g., Balsam & Payne, 1979). With this duration procedure, magnitude either has no effect on acquisition rate, or it produces faster acquisition with small rewards. As a result, timing theorists tend to assume that reinforcer magnitude does not affect acquisition or extinction. As applied to the data summarized in Table 1, this conclusion works well for acquisition data, which generally show no magnitude effects, but it does not apply to extinction results. Extinction rates are typically clearly different across groups; pigeons trained with large reinforcers extinguish relatively more slowly than pigeons trained with small reward (e.g., Papini, 1997; Papini & Thomas, 1997). This rate difference suggests that a 506 M.R. Papini et al. / Learning and Motivation 33 (2002) 485–509 transformation of extinction data in terms of expected reinforcers (cf. Gibbon et al., 1980) would not eliminate the effect of reinforcer magnitude on extinction. Timing theory offers, therefore, no clear explanation of the magnitude effects reported in spaced-trial experiments with pigeons. Also unclear is the way in which the timing hypothesis would deal with the interreinforcement interval. Exposure to the training apparatus during this interval is assumed to lead to contextual conditioning. However, in the spaced-trial procedure used in the present experiments, the pigeon was in the training context during a relatively small fraction of the interreinforcement interval; pigeons were in their cages during most of this interval. Thus, the timing hypothesis is so tied to particular training procedures and it is difficult to apply when, as in the present experiments, training involves somewhat unusual circumstances. A third theoretical possibility, referred to as the attentional hypothesis, is particularly relevant to the second dissociation mentioned previously, the different effects of partial reinforcement in between-subject versus withinsubject designs. Bouton and Sunsay (2001) suggested that partial reinforcement encourages attention to both the trial stimulus as well as the contextual cues in which trials occur because of the surprising nature of the R and N outcomes. In contrast, continuous reinforcement encourages attention only to the trial stimulus. Consistent with this attentional hypothesis, Bouton and Sunsay (2001) reported that responding to a partial stimulus is more sensitive to contextual change than responding to a continuously reinforced stimulus, in a within-subject, appetitive conditioning experiment with rats. A change from acquisition to extinction leads to a greater attentional shift away from the trial stimulus and to the contextual stimuli in the 100% reinforcement condition than in the 50% reinforcement condition, thus leading to the PREE. In a within-subject situation such as that used in Experiment 3, the presentation of the partial stimulus in the same context as the other two stimuli would tend to match the amount of attention to contextual cues across stimuli. In the absence of differential attention to the context, and applying Pearce et al.Õs (1997) rule, the S–W mechanism takes priority, producing a reversed PREE. The attentional hypothesis has not been explicitly applied to the reinforcer magnitude case. However, it seems plausible to expect that attention to the context would be inversely related to reinforcer magnitude. A testable prediction suggests that a context shift should be more detrimental to a stimulus paired with a small reinforcer than to one paired with a large reinforcer. In this case, then, the attentional hypothesis predicts that a shift to extinction should be more disrupting in the large-reinforcer group than in the small-reinforcer group (i.e., an MREE). Thus, this attentional hypothesis cannot explain the reversed bMREE (Papini, 1997; Papini & Thomas, 1997) and wMREE (Experiment 3). The results of this series of spaced-trial experiments with pigeons are difficult to explain for a wide variety of theoretical accounts, both classic M.R. Papini et al. / Learning and Motivation 33 (2002) 485–509 507 and contemporary. The merits of the response-unit, timing, and attentional hypotheses need to be established in experiments explicitly designed to test their application to the present conditions of training. A consideration of all the available evidence suggests that it may be premature to conclude that the same mechanisms explain the adjustment to reinforcement uncertainty in pigeons and rats, despite the behavioral similarities in the bPREE. The first piece of evidence is provided by the dissociation of bPREE and bMREE in spaced-trial runway and key-pecking experiments with pigeons (Papini, 1997; Papini & Thomas, 1997; Roberts et al., 1963; Thomas, 2001). Therefore, the same training conditions that yield the spaced-trial, bPREE lead to a reversed bMREE in pigeons. This dissociation of the bPREE and bMREE is inconsistent with evidence from experiments with rats, for which these effects co-vary under analogous conditions of training (Hulse, 1958; Wagner, 1961). The second piece of evidence is provided by the drug profile of the pigeonÕs bPREE, which appears to differ from that of the rat. For example, haloperidol (mainly a dopamine antagonist) eliminates the bPREE in pigeons (Thomas, 2001), but has no effect in rats (Feldon, Katz, & Weiner, 1988); by contrast, chlordiazepoxide (a benzodiazepine anxiolytic) delays the emergence of the bPREE without affecting its size in pigeons (Thomas, 2001), but it eliminates the bPREE in rats (Feldon & Gray, 1981; McNaughton, 1984). These results suggest that different underlying mechanisms contribute to the PREE in pigeons and rats. Acknowledgments Part of the results in this paper were presented at the 10th Annual Meeting of the American Psychological Society, Washington D.C., May 1998. We thank John Capaldi, Roger Mellgren, and Steven Stout for comments on a previous version of the manuscript. References Amsel, A. (1992). Frustration theory. An analysis of dispositional learning and memory. New York: Cambridge University Press. Amsel, A., Rashotte, M. E., & MacKinnon, J. R. (1966). Partial reinforcement effects within subject and between subjects. Psychological Monographs, 80(628). Anderson, N. H. (1963). Comparison of different populations: Resistance to extinction and transfer. Psychological Review, 70, 162–179. Balsam, P. D., & Payne, D. (1979). Intertrial interval and unconditioned stimulus durations in autoshaping. Animal Learning and Behavior, 7, 477–482. Boitano, J. J., & Foskett, M. D. (1968). Effects of partial reinforcement on speed of approach responses in goldfish (Carassius auratus). Psychological Report, 22, 741–744. Bouton, M. E., & Sunsay, C. (2001). Contextual control of appetitive conditioning: Influence of a contextual stimulus generated by a partial reinforcement procedure. Quarterly Journal of Experimental Psychology, 54B, 109–125. 508 M.R. Papini et al. / Learning and Motivation 33 (2002) 485–509 Capaldi, E. J. (1994). The sequential view: From rapidly fading stimulus traces to the organization of memory and the abstract concept of number. Psychonomic Bulletin and Review, 1, 156–181. Couvillon, P. A., & Bitterman, M. E. (1985). Analysis of choice in honeybees. Animal Learning and Behavior, 13, 246–254. Crawford, L. L., Steirn, J. N., & Pavlik, W. B. (1985). Within- and between-subjects partial reinforcement effects with an autoshaped response using Japanese quail (Coturnix coturnix japonica). Animal Learning and Behavior, 13, 85–92. Daly, H. B., & Daly, J. T. (1982). A mathematical model of reward and aversive I nonreward: Its application in over 30 appetitive learning situations. Journal of Experimental Psychology: General, 111, 441–480. Eisenberger, R. (1992). Learned industriousness. Psychological Review, 99, 248–267. Feldon, J., & Gray, J. A. (1981). The partial reinforcement extinction effect after treatment with chlordiazepoxide. Psychopharmacology, 73, 269–275. Feldon, J., Katz, Y., & Weiner, I. (1988). The effects of haloperidol on the partial reinforcement extinction effect (PREE): implications for neuroleptic drug action on reinforcement and nonreinforcement. Psychopharmacology, 95, 528–533. Gallistel, C. R., & Gibbon, J. (2000). Time, rate, and conditioning. Psychological Review, 107, 289–344. Gellermann, L. W. (1933). Chance orders of alternating stimuli in visual discrimination experiments. Journal of Genetic Psychology, 42, 206–208. Gibbon, J., & Balsam, P. D. (1981). Spreading association in time. In C. M. Locurto, H. S. Terrace, & J. Gibbon (Eds.), Autoshaping and conditioning theory (pp. 219–253). New York: Academic. Gibbon, J., Farrell, L., Locurto, C. M., Duncan, H. J., & Terrace, H. S. (1980). Partial reinforcement in autoshaping with pigeons. Animal Learning and Behavior, 8, 45–59. Gonzalez, R. C., & Bitterman, M. E. (1962). A further study of partial reinforcement in the turtle. Quarterly Journal of Experimental Psychology, 14, 109–112. Gonzalez, R. C., & Bitterman, M. E. (1967). Partial reinforcement effect in the goldfish as a function of amount of reward. Journal of Comparative and Physiological Psychology, 64, 163–167. Goodrich, K. P. (1959). Performance in different segments of an instrumental response chain as a function of reinforcement schedule. Journal of Experimental Psychology, 57, 57–63. Graf, C. L. (1972). Spaced-trial partial reward in the lizard. Psychonomic Science, 27, 153–154. Hulse, S. H., Jr. (1958). Amount and percentage of reinforcement and duration of goal confinement in conditioning and extinction. Journal of Experimental Psychology, 56, 48–57. McNaughton, N. (1984). Effects of anxiolytic drugs on the partial reinforcement effect in runway and Skinner box. Quarterly Journal of Experimental Psychology, 36B, 319–330. Mellgren, R. L., & Dyck, D. G. (1974). Reward magnitude in differential conditioning: effects of sequential variables in acquisition and extinction. Journal of Comparative and Physiological Psychology, 86, 1141–1148. Mellgren, R. L., & Seybert, J. A. (1973). Resistance to extinction at spaced trials using the within-subject procedure. Journal of Experimental Psychology, 100, 151–157. Mowrer, O. H., & Jones, H. M. (1945). Habit strength as a function of the pattern of reinforcement. Journal of Experimental Psychology, 43, 293–311. Muzio, R. N., Segura, E. T., & Papini, M. R. (1992). Effect of schedule and magnitude of reinforcement on instrumental learning in the toad, Bufo arenarum. Learning and Motivation, 23, 406–429. Muzio, R. N., Segura, E. T., & Papini, M. R. (1994). Learning under partial reinforcement in the toad (Bufo arenarum): Effects of lesions in the medial pallium. Behavioral and Neural Biology, 61, 36–46. Papini, M. R. (1997). Role of reinforcement in spaced-trial operant learning in pigeons (Columba livia). Journal of Comparative Psychology, 111, 275–285. M.R. Papini et al. / Learning and Motivation 33 (2002) 485–509 509 Papini, M. R. (2002). Pattern and process in the evolution of learning. Psychological Review, 109, 186–201. Papini, M. R., Muzio, R. N., & Segura, E. T. (1995). Instrumental learning in toads (Bufo arenarum): Reinforcer magnitude and the medial pallium. Brain, Behavior, and Evolution, 46, 61–71. Papini, M. R., & Thomas, B. L. (1997). Spaced-trial operant learning with purely instrumental contingencies in pigeons (Columba livia). International Journal of Comparative Psychology, 10, 128–136. Pavlik, W. B., & Carlton, P. L. (1965). A reversed partial-reinforcement effect. Journal of Experimental Psychology, 70, 417–423. Pavlik, W. B., Carlton, P. L., & Hughes, R. A. (1965). Partial reinforcement effects in a runway: Between- and within-Ss. Psychonomic Science, 3, 203–204. Pearce, J. M., Redhead, E. S., & Aydin, A. (1997). Partial reinforcement in appetitive Pavlovian conditioning with rats. Quarterly Journal of Experimental Psychology, 50B, 273–294. Pert, A., & Bitterman, M. E. (1970). Reward and learning in the turtle. Learning and Motivation, 1, 121–128. Rescorla, R. A. (1999). Within-subject partial reinforcement extinction effect in autoshaping. Quarterly Journal of Experimental Psychology, 52B, 75–87. Rescorla, R. A., & Wagner, A. R. (1972). A theory of Pavlovian conditioning variations in the effectiveness of reinforcement and nonreinforcement. In A. H. Black & W. F. Prokasy (Eds.), Classical conditioning II: Current research and theory (pp. 64–99). New York: Appleton-Century-Crofts. Roberts, W. A. (1980). Distribution of trials and intertrial retention in delayed matching to sample with pigeons. Journal of Experimental Psychology: Animal Behavior Processes, 6, 217–237. Roberts, W. A., Bullock, D. H., & Bitterman, M. E. (1963). Resistance to extinction in the pigeon after partially reinforced instrumental training under discrete-trials conditions. American Journal of Psychology, 76, 353–365. Sadler, E. W. (1968). A within- and between-subjects comparison of partial reinforcement in classical salivary conditioning. Journal of Comparative and Physiological Psychology, 66, 695–698. Schmajuk, N. A. (1997). Animal learning and cognition. A neural network approach. Cambridge, UK: Cambridge University Press. Schutz, S. L., & Bitterman, M. E. (1969). Spaced-trials partial reinforcement and resistance to extinction in the goldfish. Journal of Comparative and Physiological Psychology, 68, 126–128. Schwartz, B. (1982). Interval and ratio reinforcement of a complex sequential operant in pigeons. Journal of the Experimental Analysis of Behavior, 37, 349–357. Sheffield, V. F. (1949). Extinction as a function of partial reinforcement and distribution of practice. Journal of Experimental Psychology, 39, 511–526. Stout, S. C., Muzio, R. N., Boughner, R., & Papini, M. R. (2002). After effects of the surprising presentation and omission of appetitive reinforcers on key pecking performance in pigeons. Journal of Experimental Psychology: Animal Behavior Processes, 28, 242–256. Svartdal, F. (2000). Persistence during extinction: Conventional and reversed PREE under multiple schedules. Learning and Motivation, 31, 21–40. Thomas, B. L. (2001). Determinants of the spaced-trial partial reinforcement extinction effect in pigeons. Unpublished Doctoral Dissertation, Texas Christian University. Thorndike, E. L. (1911). Animal intelligence. New York: Macmillan. Wagner, A. R. (1961). Effects of amount and percentage of reinforcement and number of acquisition trials on conditioning and extinction. Journal of Experimental Psychology, 62, 234–242. Weinstock, S. (1954). Resistance to extinction of a running response following partial reinforcement under widely spaced trials. Journal of Comparative and Physiological Psychology, 47, 318–322.