Delay of reinforceme..

advertisement
A Quantitative Analysis of the Behavior Maintained by Delayed Reinforcers
A. Charles Catania
University of Maryland, Baltimore County (UMBC)
Mark P. Reilly, Dennis Hand
Central Michigan University
Lara Kowalsky Kehle, Leanne Valentine, Eliot Shimoff
University of Maryland, Baltimore County (UMBC)
Suggested Running Head: Behavior Maintained by Delayed Reinforcers
Corresponding Author
A. Charles Catania
10545 Rivulet Row
Columbia, MD 21044
Phone: 410-730-6949
FAX: Same as phone, but requires advance notice
Email: catania@umbc.edu
Catania et al., Page 2
Abstract
Random-interval reinforcement contingencies were arranged for a sequence of pigeon pecks
on one key followed by a sequence on a second key. Whether the second sequence required
specified numbers of pecks, minimum times preceding reinforced pecks, or both, first-key
pecking was separated from reinforcers by delays that included number of pecks and time. Firstkey response rates decreased as number of responses or time separating them from reinforced
second-key responses increased. These functions, delay gradients, were obtained in one
experiment in which reinforced sequences consisted of M first-key pecks followed by N secondkey pecks, where M+N equaled 16, in a second in which the number of first-key pecks was held
constant at 8, and in a third that varied the minimum delay between most recent first-key pecks
and reinforced second-key pecks. In each case, gradients were equally well fit by exponential,
hyperbolic and logarithmic functions. Performances were insensitive to reinforcer duration, and
functions were consistent across varied random-interval values. In one more experiment,
differential reinforcement of rate of second-key responding during delay intervals independently
varied time and number. Forms of the delay gradients depended primarily on time rather than on
number of second-key responses separating first-key responding from the reinforcer.
(199 words)
Key words: Delay of reinforcement, delay gradient, time versus number, random-interval
schedule, topographical tagging, key pecks, pigeons
Catania et al., Page 3
Delay of reinforcement has a long history in research on learning, whether it has been
evoked in explicit terms (e.g., Hamilton, 1929; Perin, 1943; Renner, 1964; Seward, 1942;
Watson, 1917; Wolfe, 1934) or only implicitly (e.g, the goal gradients of Hull, 1943).
Throughout much of its history, delayed reinforcement has been regarded mainly as an
impediment to the acquisition and maintenance of behavior rather than as a ubiquitous property
of environments that can organize extended behavior sequences. Other perennial issues are what
fills the time between a reinforced response and its later reinforcing consequence and the
functional properties of what fills it (e.g., Keller & Schoenfeld, 1950; Laties, Weiss, Clark, &
Reynolds, 1965). The idea that intervening responding matters has a long history, as in the
following early example: “It may be that time in delay of reward is important only in so far as it
is roughly proportional to the amount of activity or the number of responses which occur
between the response to be strengthened and the occurrence of the reinforcement” (Grice, 1942,
p. 488). But we will find that elapsed time is a more important determinant of the effects of
delayed reinforcers than quantity of intervening behavior.
If two events are simultaneous, one of course cannot have been the cause of the other, so
responses are necessarily followed by their consequences only after some delay. Considerations
of how to deal with the gap have received substantial attention, but while we acknowledge that
behavior is necessarily extended in time, what bridges the gap will not be our main concern here.
The issue is not about molar versus molecular analyses; even so-called molecular measures such
as interresponse times (IRT) are based on distributions obtained over extended time periods (cf.
Baum, 2004; Hineline, 2001). We will also have little to say about delay as it arises in
respondent procedures (Pavlov, 1927). Our main objective is to characterize the function relating
the rate of a response to the delay that separates it from a subsequent reinforcer, to establish the
Catania et al., Page 4
appropriate dimensions for such a functional relation, and to determine some of the variables that
may serve as its important parameters (cf. Lattal, 1984, 2010).
A milestone in the analysis of delay came with the work of Dews (1960, 1962, 1966, 1970),
especially with his discovery that the fixed-interval (FI) scallop, the increasing response rate
within each interval, survives repeated interruptions. The phenomenon is illustrated in Figure 1.
Because this pattern could not plausibly be accounted for in terms of discriminative effects of
current behavior or other appeals to properties of responding, Dews concluded that the FI scallop
emerges because the reinforcer at the end of the interval has effects that depend on its relation to
all of the preceding responses and not just the one that produced it. Contingent on the time
elapsed since the start of the interval, that reinforcer strengthens responding based on its relation
to each response in the interval. FI scallop as the joint product of a gradient of temporal
discrimination, in that the organism’s different rates at different times in the interval imply its
discrimination of elapsed time, and a gradient of delayed reinforcement, in that responses at
various locations in the interval are consistently followed by the reinforcer at particular delays
(e.g., in FI 500 s, responding at 400 s is typically followed not less than 100 s later by a
reinforced peck). The two gradients must exist together, but it is important to recognize that
responding at any time in an interval depends on what happened after that time in previous
intervals; it cannot depend on the reinforcer at the end of the current interval, because that
reinforcer has not been delivered yet (see also Church & Gibbon, 1982; Neuringer, 1969).
<Insert Figure 1 about here>
In the early days of reinforcement schedules, schedule effects were discussed as partial
reinforcement, and it was seen as a paradox that more behavior could be generated by reinforcing
some fraction of the total responses than by reinforcing every response. But looking at schedules
Catania et al., Page 5
in terms of the delayed reinforcement of all responses that precede the reinforced response
suggests that intermittent or partial reinforcement works as it does because it allows each
reinforcer to reinforce many responses instead of just one. Figure 2 makes the point
schematically. Both 1 and 2 show a time sample including an early response, b, and a later
reinforced response, a. In 1, no other responses intervene; in 2, responses c through j occur
between them. One way to interpret these two cases is to say that in 1 the second of two
responses was reinforced (1/2) whereas in 2 only the last of ten was reinforced (1/10). But that
interpretation assumes that the effect of the reinforcer was limited to a even though it followed
the other responses too.
<Insert Figure 2 about here>
An alternative, following from Dews’ insight, is to assume that the effect of the reinforcer
depends on its relation to all of the responses that preceded it, some after longer delays than
others. From that perspective, only two responses were reinforced in 1 whereas ten were
reinforced in 2. In this view, the partial reinforcement effect need not be regarded as paradoxical.
Instead, schedules allow each reinforcer to act on several responses at a time instead of just the
most recent one.
Although the earlier responses in a sequence that ends with a reinforcer contribute to future
responding based on their relation to the reinforcer, they contribute less than the later ones
because of the longer delays that separate them from the reinforcer (Dews, 1962). This means
that in interpreting effects of schedules, it would help to know the form of the delay gradient.
The delay gradient has entered successfully into mathematical models of operant behavior (e.g.,
Berg & McDowell, 2011; Catania, 2005e, 2011b; Killeen, 1994), but technical problems
complicate its experimental determination. For example, as illustrated in Figure 3, if we arrange
Catania et al., Page 6
delayed reinforcers only some fixed time after a response, we must either allow additional
responses to occur in the interim, in which case the actual time from the last response to the
reinforcer will typically be shorter than the one we scheduled, or we can reset the delay with
each subsequent response, in which case the resetting of the timer will differentially reinforce
pauses at least as long as the delay interval, and that differentiation will be confounded with the
delayed effects of the reinforcer (e.g., Catania & Keller, 1981; Katz & Catania, 2005). We
cannot avoid these problems by presenting a stimulus during which the delay operates, because
that simply substitutes an immediate conditional reinforcer for the delayed one.
<Insert Figure 3 about here>
Delays can be discussed in terms of time elapsed since a response, but they can also be
interpreted in terms of decaying traces of events correlated with responses (e.g., Killeen, 1994,
2001, 2011; Killeen & Sitomer, 2003). Accounts in terms of delay between a response and a
reinforcer and decay of memory for a response can be made quantitatively equivalent, but until
the time at which decay can be identified with physiological processes, we will here favor a
vocabulary in terms of dimensions that can be manipulated directly in experimental procedures
(cf. Wixted & Gaitan, 2002). Time between a response and a reinforcer is one such dimension
(cf. Catania, 1970, 1991).
Figure 4 shows hypothetical gradients that increase as the time to the reinforcer decreases.
In the top example, only one response occurs, just before the reinforcer, and it makes a
contribution to subsequent responding proportional to the height of the gradient at that single
point. In the bottom example, eight responses occur and, based upon Dews’ arguments, we will
assume that each response contributes to subsequent responding proportional to the height of the
gradient at the time at which it occurs. In other words, consistent with what has been well-
Catania et al., Page 7
established in the literature on delay of reinforcement, the shorter the time to the reinforcer, the
bigger its effect on subsequent responding.
<Insert Figure 4 about here>
One method for studying delay gradients is topographical tagging, arranging contingencies
so that responses followed by particular delays are distinguished topographically, as, for
example, by location (Catania, 1971). The procedure is illustrated in Figures 5 and 6. In a singlekey procedure, as shown in Figure 5, the response sequence ba produces a reinforcer. In a twokey procedure, the ba sequence again produces a reinforcer, but this time a contingency is
arranged so that b occurs on Key B rather than Key A. It is reasonable to assume that subsequent
responding on Keys A and B depends on the respective contributions of the relation of response
a to the reinforcer and that of response b to the reinforcer. Figure 6 shows a sequence of
responding in which contingencies have produced a sequence of responses on one key (top)
followed by a sequence on a second key that ends with a reinforced response (bottom). If future
responding on the first key is determined by the summed delay gradient heights at the time of
each response to the left of the dashed line, the total output on the two keys would be expected to
be about the same as the output if all responses had occurred on the second key, because the
reinforcer would have provided the same cumulative reinforcing effect in each case.
<Insert Figures 5 and 6 about here>
When the separation of a Key-B response from the reinforcer was increased by requiring
longer sequences of pecks on Key A (e.g., baa, baaa, baaaa, etc.), the rate of Key-B responding
decreased, as illustrated in Figure 7, which plots the Key-B response rate against the average
time (sum of interresponse times or IRTs) separating the Key-B response from the reinforcer
(derived from data presented in Catania, 1971). In this procedure, all reinforcers are produced by
Catania et al., Page 8
a Key-A peck at the end of the sequence, so relative Key-A and Key-B rates can be determined
only by relative differences in delay and not by relative reinforcement.
<Insert Figure 7 about here>
Proposals for the form of the delay function have most prominently included exponential
functions, logarithmic functions and variations on hyperbolic functions (Catania, 2005e; Chung,
1965; Killeen, 2011; Mazur, 1987, 1995). Choosing among them raises several problems, not
least of which is the measure of responding. For example. we may choose response rate or
resistance to change (Nevin & Grace, 2000; Podlesnik, Jimenez-Gomez, Ward, & Shahan, 2006;
Shahan & Lattal, 2005), and if we study response rate, as in the present research, procedures may
dictate the options. Figure 8 provides an illustration. A procedure that estimates indifference
points with different delays pitted against each other, as in Mazur (1987, 1995), yields the height
of the function at a given delay (top gradient in Figure 8); one that examines responding over a
range of time, as in topographical tagging, yields an area under the function (bottom gradient of
Figure 8). But if the former is hyperbolic then the latter should approach a logarithmic function
(Killeen, 1994). Furthermore, if the parameters of exponential functions are themselves variable,
they can yield data best fit by power functions, of which one is the hyperbolic (Murre & Chessa,
2011). Although we will favor exponential fits to our data, the differences in variances accounted
for by these and other functions are typically too small to favor one strongly over another.
<Insert Figure 8 about here>
One more preliminary remains before we can proceed to the research proper. In dealing with
delay gradients, we must determine how far back from the reinforcer we should extend them. A
body of evidence strongly implies that for any given reinforcer its effect is limited to responses
that occur during the time that separates it from the previous reinforcer (Catania, Sagvolden, &
Catania et al., Page 9
Keller, 1988), as illustrated in Figure 9. The top gradient shows the hypothetical effect of a
reinforcer, rfrx, produced by a response at the end of some time period, where the effect of the
reinforcer on a response is given by the height of the gradient at the time at which it occurs. The
bottom gradients show the effect of an earlier reinforcer, rfry, on the rfrx gradient; rfry generates
its own gradient, but it truncates the tail of the rfrx gradient. The implication is that the
introduction of rfry reduces the responding maintained by rfrx, and this is of course consistent
with what we know about the interacting effects of reinforcers (e.g., Catania, 1969, 1973).
<Insert Figure 9 about here>
The effect in the figure can be thought of as a blocking effect, and it is not clear whether it
would be attenuated with varied instead of constant reinforcers (e.g., Steinman, 1968), or if some
reinforcers were delivered independently of responding (cf. the concept of coupling, as in
Killeen & Sitomer, 2003). This feature of delay gradients was an essential component of a
mathematical model that assumed that, as in Skinner’s Reflex Reserve (Killeen, 1988; Skinner,
1938, 1940), all of the responses impacted by the delay gradient contributed to a reserve of
response potential that was depleted as subsequent responses were emitted (Catania, 2005e).
General Method
The experiments that follow are variations on topographical tagging procedures conducted
with pigeons either in the pigeon laboratories of the Department of Psychology at the University
of Maryland, Baltimore County (UMBC) or those of the Department of Psychology at Central
Michigan University (CMU). We begin with some features common to all or most procedures,
with details to follow later.
Catania et al., Page 10
Subjects
White Carneaux or Silver King pigeons with histories in various operant procedures served
in the research conducted at UMBC. Homing Pigeons provided with delayed-reinforcement
histories in preliminary sessions served in the research conducted at CMU. All pigeons were
housed in individual cages with free access to water in a room with a 12-hr light/dark cycle, were
maintained at approximately 80% of free-feeding weights, and were provided supplemental
feeding to maintain body weights after each session. Experimental sessions were conducted in
the light portion of the light/day cycle. All pigeons were treated in accordance with the
respective UMBC and CMU Institutional Animal Care and Use Committee guidelines, and
procedures were conducted in accordance with protocols approved by those committees. Upon
the closing of the UMBC laboratories (see Catania, 2005c), all pigeons housed there were placed
with a local pigeon fancier.
Apparatus
Daily sessions were conducted in pigeon chambers that included two active keys, a feeder,
and a houselight. In the UMBC pigeon laboratory, all chambers included two standard keys
separated 6.4 cm center-to-center 28 cm above the floor. Below the keys was a 10-cm-square
opening to a standard Gerbrands feeder that provided fixed durations of access to Purina
Nutrabrand pellets as reinforcers. Reinforcer durations were adjusted for each pigeon so as to
minimize postsession feeding, typically within the range from 3 to 5 s; they remained constant
throughout an experiment for each pigeon, except for Experiment 5 on reinforcer duration
Chambers were controlled by Alpha-Bus interfaces operated by an assembly language
routine nested within FORTH programs (see Appendix I in Catania, 2005e) and running on PS/2
or other IBM computers with MS-DOS operating systems. These arrangements did not allow for
Catania et al., Page 11
the collection of individual responses throughout sessions, but a variety of detailed measures
summated over individual sessions were recorded for each pigeon, some of which are provided
in Appendixes 1 and 2.
Sessions at CMU were conducted in four equally equipped Med-Associates (St. Albans,
VT) pigeon chambers, each 27 cm wide by 28 cm tall by 25 cm deep and each containing three
2.5-cm-diameter pigeon keys 21 cm above the floor and 8 cm apart. Floors consisted of steel
rods spaced approximately 1 cm apart suspended over a waste tray. The keys were located above
a 5.7 cm tall by 5.2 cm wide opening behind which was mounted a solenoid-operated Coulbourn
Instruments (Allentown, PA) feeder, which provided white millet and was illuminated by a 27-v
white light during reinforcer deliveries. The right key and two 4.5-cm-wide levers mounted
approximately 1 cm above the floor and directly below the right and left response keys were not
used. Only the left and center keys were used, both lit red in the main procedure, but for
consistency with discussions of data from two-key chambers, they will subsequently be referred
to as the left and right keys. A 25-v houselight was on the opposite wall. Computer control
allowed all responses to be recorded, so that data analysis was not limited to summary measures
decided upon in advance.
Hall-Effect Pigeon Keys. Instead of the electrical make-break contacts of standard
Gerbrands pigeon keys, the keys in the UMBC experiments relied on the Hall effect, the
production of current by the movement of a magnet relative to a sensor, and therefore obviated
concerns about changes in conductivity due to arcing at key contacts. As shown in Fig. 10, a
magnet attached to the moving portion of the key rested against the Hall-effect switch
(Multiplexed two-wire Hall Effect sensor IC, Model UGN3055U, from Allegro Microsystems,
Worcester, MA). Pecks moved the magnet away from the switch and generated an output
Catania et al., Page 12
processed at the Alpha-Bus interface. Switch reset as the magnet returned to its resting position
involved sufficient hysteresis that key bounce or chatter was not a significant problem. Key
tension, set to record pecks of about 1.5 N or greater, was adjusted in the usual way and
maintenance consisted mainly of minor cleaning so that pigeon dust or other debris did not build
up between the magnet and the switch. We found the keys to be highly reliable, and calibration
by Shimoff via the assembly portion of the program showed them capable of following response
rates up to roughly 20 responses/s.
<Insert Figure 10 about here>
Procedure
In all procedures, random-interval (RI) contingencies were arranged for sequences of pecks
on each of two pigeon keys. In one arrangement, for example, with an RI schedule of 60 s,
reinforcers became available (were set up) on the average of once a minute for the last peck of a
sequence of exactly M left-key pecks followed by exactly N right-key pecks. Figure 10
illustrates the procedure schematically and provides some nomenclature.
<Insert Figure 11 about here>
Data Analysis. The figure shows a hypothetical sequence of responses on left (L) and right
(R) keys as they might occur between two reinforcer deliveries (RFR) in a well-established
performance in which the sequence to be reinforced is exactly eight left pecks followed by
exactly eight right pecks. Pecks occur as runs of responses on one key uninterrupted by
responses on the other key (L RUNs and R RUNs) and runs are separated by changeovers from
left to right (LR) or from right to left (RL). A sequence consisting of an L RUN, an LR
changeover, an R RUN and an RL changeover constitutes a CYCLE.
Catania et al., Page 13
To insure homogeneity of the data entering into computations of response rates, some
portions of the time between two successive reinforcers were excluded from the analysis. These
included post-reinforcer pauses (PRPs) and all cycles that ended with a reinforced right-key
peck. These data-analysis boundaries are represented by the dashed verticals at a and b in the
figure. Post-reinforcer pauses (PRP) were excluded because they were typically longer than IRTs
and changeovers. In a history of pilot sessions with these procedures, all pigeons came to
respond almost exclusively on the left key after each PRP, and this pattern remained in place
throughout all subsequent sessions, so no provision was made for the very rare right-key
responses that occurred after a PRP. Also, for the purposes of data analysis the time to the first
peck of the session was treated as equivalent to a PRP. These exclusions were for the purposes of
data analysis only: the clocks for overall session duration and for the scheduling of RI setups
stopped only during operations of the pigeon feeder.
Correcting for Selective Stopping. Cycles ending with a reinforcer were excluded because
they constituted the selective stopping of response sequences, with the potential consequence
that the required L and R pecks might be overrepresented in derived measures such as L and R
response rates. Consider sequences of three responses, each l or r, with responding terminated
only with the reinforcement of two successive r responses. If the responses occur with roughly
equal frequencies, the possible sequences are lll, llb, lrl, lrr, rll, rlr, rrl and rrr, but if two
successive reinforced r responses terminate a sequence, the behavior recorded will consist of lll,
llr, lrl, lrr, rll, rlr, rr and rr and the sequences ending with a reinforcer will consist of lrr, rr and
rr, which include only a single l among seven responses. If the sequences not ending with a
reinforcer continue, they again end only when they include two successive r responses, and so
on, thereby maintaining the bias toward higher r than l response rates produced by the selective
Catania et al., Page 14
stopping. Overall response rates for a given key were therefore based upon total responses over
the cumulated time from the end of the PRP to the start of but excluding the reinforced cycle (a
to b in Figure 10). These overall rates are to be distinguished from local rates, estimated by
dividing mean responses per run by mean run duration.
Random-Interval Scheduling. Contemporary arrangements usually program the variety of
variable-interval schedule called random interval (RI), in which the repeated sampling of a
probability gate produces reinforcement setups. For example, producing a setup with a
probability (p) of .02 at the end of every second (t) arranges a mean interval of t/p, or 50 s. In the
days of electromechanical equipment, intervals were instead scheduled by a motor-driven loop of
tape with holes punched in it, with minimum intervals dictated by the minimum spacing that
allowed a switch to throw completely as its sensor cleared one hole and then fell into the next
(e.g., Catania & Reynolds, 1968; Ferster & Skinner, 1957). This led to the practice of stopping
the programming tape each time a punched hole was sensed. The tape started again only after the
set up reinforcer was delivered. Without this arrangement, the next hole in the tape could move
past the sensor without detection, thereby losing reinforcers that should have been scheduled.
But stopping the tape also added time to intervals, so obtained rates of reinforcement were
typically lower than scheduled ones.
. Contemporary interval programs usually preserve the procedural artifact, though it arose
from the limitations of electromechanical equipment. To stop probability sampling once a
reinforcer has been set up in an RI schedule is functionally equivalent to stopping the tape when
setups are arranged electromechanically. But in computer programs setups need not be treated as
yes-no states; instead, their value can be incremented with each setup and decremented with each
reinforcer delivery, and probability sampling can continue even if reinforcers already set up have
Catania et al., Page 15
not yet been collected. With this arrangement, for example, if three setups have occurred but
responding has briefly ceased for some reason, the next three responses in a row will be
reinforced once responding resumes (cf. Catania, 2002, 2011b).
Figure 12 compares traditional RI scheduling and this alternative, which was used for the RI
schedules in the present research. Because reinforcer durations were excluded from calculations
of session duration, reinforcers have been represented as essentially instantaneous. The top
example illustrates standard practice: Scheduling stops with each setup until a response produces
the current reinforcer. No setups are lost, but the time from setup to reinforced response adds to
each scheduled interval. In the example only 16 of the 18 reinforcers that could have been
available are actually delivered; two that would have been delivered if timing had not stopped
during setups have been pushed beyond the available time..
<Insert Figure 12 about here>
In the bottom example, timing continues after each setup and setups accumulate. Thus,
when setups 7 and 8 occur during a single IRT, the next two responses each produce a reinforcer.
After four setups during a later long pause, four responses in a row produce reinforcers. By
session end, all 18 scheduled reinforcers have been delivered. Occasional setups can be lost if a
session ends without a response just after they are arranged. Even so, these contingencies
produce much closer correspondences between scheduled and obtained reinforcers than those in
the traditional example. (This might not be the case with schedules in which the effect of the
reinforcer may oppose differential reinforcement, as in tandem RI differential-reinforcement-oflow-rate or drl schedules).
Furthermore, the number of uncollected setups remaining at the end of a session is an
index of control by the scheduled contingencies, especially where these involve the differential
Catania et al., Page 16
reinforcement of complex response sequences. If very few setups remain uncollected at the end
of a session, this is evidence that responding has conformed to the schedule contingencies, so
this number should rarely be larger than zero or one in steady-state performance. In our
procedures, therefore, the number of uncollected setups at the end of a session was used as a
measure of stability in our decisions about whether to advance to a new experimental condition.
The topographical tagging schedule. Unless otherwise noted, an RI 60-s schedule operated
to set up reinforcers for the last of a sequence of exactly M left-key pecks followed by exactly N
right-key pecks, daily sessions were arranged for a maximum of 60 min or 60 reinforcers,
whichever came first, and reinforcer duration was 4 s, with reinforcer time excluded from the
timing of session duration. Arranging a reinforcer for the last peck of the M+N sequence meant
that a sequence that had already been initiated could produce a reinforcer if a setup occurred
during its emission. Without this contingency, interreinforcer intervals would always have been
lengthened by the time it took for the sequence to be completed, so that procedures that required
longer sequences would also have produced lower reinforcement rates. It is tempting to describe
the contingencies as a tandem schedule that includes the RI along with the two FR components,
but the setup contingency makes it more appropriate instead to describe it as RI 60 s in which the
higher-order reinforced operant consists of the tandem sequence FR M (left key) FR N (right
key), or, in other words, as an RI 60-s (tandem FR M FR N) schedule. We will often refer to M
and N as the left-key ratio and the right-key ratio.
Experiment 1: Varying Delay by Varying Number of Intervening Responses (Part I)
The procedures that provided the data in Figure 8 were obtained while M was held constant
at 1 and N varied (Catania, 1971). The variations in N effectively added a varying fixed-ratio
Catania et al., Page 17
contingency to right-key responding. The present experiment changed the delayed reinforcement
of left-key responding by varying N also, but to avoid varying the overall ratio contingency it
held the sum of M plus N constant at sixteen. Another assumption was that responses that were
seventeen or more responses back from the reinforcer were far enough removed that their
cumulative effects were small compared to those for the sixteen for which contingencies were
arranged. But if delays are implicated even in the early parts of long FI schedules, then certainly
they can operate beyond the times occupied by the required response sequences, so this is an
issue that will probably needed to be revisited. This was beyond the reach of the present analyses
because most data were collected by computers with insufficient capacity to record entire
sessions in detail. Overall response rates may allow the proportions of lefts and rights in such
cases to be estimated but we have not attempted to do so.
Method
The subjects and apparatus were as described in General Method. All pigeons were White
Carneaux; Pigeons 73, 74 and 76 were males and Pigeon 78 was a female. The sessions were
preceded by several weeks of pilot sessions during which equipment and programs were
calibrated and tested, the pigeons were provided with a history of different M and N values, and
their steady-state performances and transitions to new parameter values were observed to insure
that each was a suitable preparation in terms of key-peck topography and variability for the
purposes of the study (see Methodological Note in Catania, de Souza, & Ono, 2004; see also
Staddon, 1965, on metastability).
Conditions included a range of values from M=1 and N=15 to M=15 and N=1, as outlined in
Appendix 1; data are means across the last five sessions of each condition. Most conditions
lasted fourteen daily sessions and were extended if performances showed two or more
Catania et al., Page 18
uncollected setups at the end of a session; this criterion was supplemented by visual inspection of
the data, where sessions were also extended if session-to-session variability was apparent even if
the setup criterion was met. Stability was judged independently for each pigeon. Because of staff
availability on weekends and holidays, experimenter error, apparatus problems or other issues,
on a few occasions fewer than fourteen sessions were arranged for a condition or a condition
continued even though stability criteria had been met. Subsequent data inspection showed that
except at the extremes of parameter values pigeons’ behavior adjusted to new conditions quickly
after transitions, usually within just one or two sessions (cf. Reilly & Lattal, 2004).
Results
Figure 13 shows response rates on the left key for each pigeon as a function of the mean
time to a reinforcer, i.e., the time separating the last left-key peck from the next reinforced rightkey peck, with exponential fits to the data for each bird (see Experiment 3, Discussion, for the
details of these and other fits). One assumption in topographical tagging is that the procedure
involves the movement of responses attributed to certain ranges of delays from one key to
another without affecting total output. Figure 14 shows the sum of the response rates across the
two keys. The linear trend-lines for each bird show overall rates decreasing with required left
responses for Bird 73 and increasing for Bird 76, with the slope essentially flat for the other two.
Given the opposite directions of effect for Birds 73 and 76, the overall data are therefore
reasonably consistent with the assumption, especially given that the different proportions of ratio
contingencies across the two keys at different parameter values might have been expected to
matter.
<Insert Figures 13 and 14 about here>
Catania et al., Page 19
Figure 15 shows that the responses per run for both the left key and the right key closely
followed the required responses on those keys (M and N). Lines show linear fits, with the leftkey slopes -1.00, -1.15, -0.81, and -0.99 and the right-key slopes 1.19, 1.21, 1.29 and 1.32 for
Pigeons 73, 74, 76 and 78 respectively. A corresponding match to the contingencies did not
appear in the delays produced by the different values of N (required right-key pecks). Figure 16
shows the mean time from the last left peck to the next reinforced right peck as a function of N.
The relation is nonlinear and is shown with an exponential fit.
<Insert Figures 15 and 16 about here>
Other measures of performance are shown in Figure 17. Changeovers from left to right were
characteristically short and did not vary systematically with parameter values across pigeons.
Consistent with the ratio contingencies created by the M and N requirements, changeovers from
right to left were of substantial duration, as were postreinforcement pauses except for those of
Pigeon 76. Again, changes with parameter values were inconsistent across pigeons, as shown by
the slopes of the linear fits for changeovers (in the aid of clarity, no fit is shown for
postreinforcer pauses, unfilled circles).
<Insert Figure 17 about here>
The overall performance is summarized in Figure 18, which includes both left and right
response rates as a function of the delay created by N, the right-key requirement. The left-key
rates are well-fitted by exponential functions, but not the right-key rates, for which logarithmic
functions have been fitted instead. Consistent with the contingencies, the rates on the two keys
are inversely related, i.e., left-key responding increases as right-key responding decreases, and
vice versa.
<Insert Figure 18 about here>
Catania et al., Page 20
Discussion
The main question, of course, is whether it is appropriate to regard the left-key data as
representing a delay gradient. Clearly left-key rates decrease as the separation between left-key
rates and later reinforcers grows. Presumably part of the decrease occurs because that separation
is also correlated with shorter required sequences on the left key. A brief exploratory study with
Pigeons 74 and 78 that examined some parts of the M and N contingency space provided some
relevant data. With other variables maintained as in Experiment 1, N was held constant at 8
while M was set at 8, 10, 12, 4 and 2 for 7, 7, 7, 14 and 9 sessions respectively. Figure 19 shows
that left-key response rates increased with increasing values of M for both pigeons.
<Insert Figure 19 about here>
We have taken the position that the delay gradient is a fundamental property of reinforced
behavior and that other properties of the performance, such as the approximate matching of
responses per run to the respective left (M) and right (N) requirements, are indirect products of
this primary contingency as it interacts with differential reinforcement operating over extended
periods of time. We do not appeal to contiguity, which in our view has been out of favor in
accounts of operant performances maintained by schedules for half a century or more (Catania,
2012; Dews, 1962; Staddon & Simmelhag, 1971). Those who begin with the assumption that
allocation and matching is primary (Baum, 2012), with other details dropping out as a
consequence of their interaction with other processes, will presumably disagree. One problem we
have with accounts in terms of allocation based on matching is that they force all responding into
a mold in which the sum of all behavior equals one, thereby precluding the possibility that the
total behavior of an organism can vary in quantity (cf. Catania, 1983).
Catania et al., Page 21
Experiment 2: Varying Delay by Varying Number of Intervening Responses (Part II)
Whichever position is held, one obvious next step is to examine gradients obtained when
required left-key responses (M) do not vary. Experiment 2 held M constant and varied N, the
required right-key responses leading up to the reinforcer.
Method
The same pigeons and apparatus were used as in Experiment 1. The primary difference in
procedure is in the required M and N responses of the higher-order RI 60 s (tandem FR M FR N)
schedule. This time, M was held constant at 8 while N was varied. Order of conditions, sessions
and other details are summarized in Appendix 2.
Results
With these parameter values, the time from the last left peck to the next reinforced right
peck changed roughly linearly with required right pecks, as shown in Figure 20. This implies that
right-key responding (changeovers from left to right and local right-key rates during runs)
remained fairly stable across experimental conditions. Figure 21 shows left-key rates as a
function of delay, i.e., the time from the last left peck to the next reinforced right peck. Like the
functions in Figure 13, all are reasonably fitted by exponential functions, though on the whole
they are less steep than the earlier functions.
<Insert Figures 20 and 21 about here>
The combined left plus right response rates and the right-key response rate alone are shown
in Figure 22. Despite the larger M+N ratio contingencies, from 8+1 = 9 to as high as 8+16 = 24
for Pigeon 76, overall rates did not change consistently, decreasing for Pigeon 73, increasing for
Pigeons 74 and 76, and increasing only very slightly for Pigeon 78. Although these data are not
Catania et al., Page 22
completely consistent with the assumptions of the topographical tagging procedure, given the
changing contingencies across conditions neither do they contradict those assumptions.
<Insert Figure 22 about here>
Right-key response rates increased with increasing delays for left-key responding and the
correlated decreases in left-key response rates, but as in Experiment 1 these data were not well
fitted by exponential functions; the fits shown are logarithmic.
Discussion
Right-key responding fills the time between left-key pecks and the later reinforcer, but a
return to the left key before the right-key sequence has been completed resets the sequence, just
as an early response resets the delay in a single-response resetting-delay procedure (cf. Figure 3).
The topographical tagging procedure substitutes alternative behavior for the period of no
responding in the resetting delay procedure, but simply providing an alternative response cannot
eliminate the reset contingency. On the other hand, if the assumptions of topographical tagging
are appropriate, the responding that becomes the alternative behavior stands in the same relation
to later reinforcers as it would have in single-key performances. Thus, it is probably still
reasonable to regard left-key responding as just the amount of behavior that can be maintained
by a given delay between responding and a later reinforcer. In other words, if there is some
maximum quantity of behavior that can be supported by a given delay of reinforcement, then our
data should at least approximate that limit, thereby perhaps narrowing the range of functions that
can describe the relation between responding and reinforcer delays.
Catania et al., Page 23
Experiment 3: Varying Delay by Varying Intervening Time
The procedures of Experiments 1 and 2 specified numbers of responses (N) intervening
between the last left peck and the next reinforced right peck. Actual delays were determined by
the rate of the intervening right responses (cf. Figures 16 and 20). Another approach is to arrange
a minimum time between the last left peck and the next reinforced right peck. In this case, a
minimum time is specified and the number of intervening right responses may vary. Analogous
to the previous RI 60-s (tandem FR M FR N) schedules, therefore, this experiment arranged RI
60-s (tandem FR M FI T-s) schedules, or, in other words, the RI 60-s reinforcement of a higherorder succession of FR M and FI T-s components, where the FR determines the required number
of left pecks and the FI determines the minimum time after which a left response can be followed
by a later reinforced right response.
Method
Subjects and Apparatus. Three male Homing Pigeons, 11, 54, and 64, served in daily
sessions in the CMU laboratory. A fourth pigeon with problems of variability and topography
was dropped as being an unsuitable preparation for the study (cf. Methodological Note in
Catania et al., 2004). The sessions were arranged in a three-key pigeon chamber in which only
two keys were functional (cf. General Method).
Procedure. Reinforcers were always 3 s of access to white millet. After some preliminary
sessions, training included tandem schedules in which reinforcers were contingent upon FR 1 for
a left-key peck followed by FR 1 for a right-key peck, with both keys lit red. Subsequently,
contingencies were gradually modified, via tandem FR 1 (left) FI 1-s (right) schedules, tandem
FR 4 (left) FI 0.5-s (right) schedules in which the FR reset if a right-key peck occurred before the
ratio was completed, and then an RI 2.5-s (tandem FR 4 FI 0.5-s) schedule. Schedule parameters
Catania et al., Page 24
were changed over roughly the next two weeks of sessions until they had reached RI 60-s
(tandem FR 4 FI 0.25-s), at which point the experiment proper began and during which sessions
lasted until 50 reinforcers had been delivered or 50 minutes had elapsed, whichever came first.
As in Experiments 1 and 2, calculations of left-key and right-key response rates excluded
reinforced sequences and postreinforcement pauses (cf. Figure 11).
For all pigeons, FI values of 0.25, 0.50, 1.00, 2.00 and 4.00 s were arranged in ascending
order. Each value was maintained for a minimum of twenty sessions, with stability then judged
based on the most recent six sessions. Stability criteria included two components: no upward or
downward trend in left-key response rates could be evident across the six sessions; and the
difference between the mean response rate over the last three sessions and over the preceding
three sessions could not exceed 5% of the mean for all six sessions. Once both criteria were
satisfied, the pigeon was advanced to the next FI value. Sessions and data are summarized in
Table 1.
<Insert Table 1 about here>
Results
Figure 23 shows the rate of left-key responding as a function of the time between the last
left-key peck and the next reinforced right-key peck for each pigeon, with data fitted by
exponential functions. As shown in Figure 24, right-key response rates tended to increase with
increasing delay for all three pigeons, whereas combined left-key and right-key rates decreased.
Given the variability and the limited number of data points, no fits are shown. As in Experiment
2, the actual delay from the last left peck to the next reinforced right peck increased roughly
linearly with the scheduled delay contingency, which in this case was the value of the terminal FI
schedule (see Table 1).
Catania et al., Page 25
<Insert Figures 23 and 24 about here>
Discussion
In judging whether the functions in Figure 23 should be interpreted as delay-ofreinforcement gradients, it might be useful to characterize them mathematically. We have shown
exponential fits in our data presentations primarily for reasons of precedent and expedience.
Table 2 provides least-squares fits of exponential, hyperbolic and logarithmic functions to the
delay data of Experiments 1, 2 and 3. These functions and their variations do not exhaust the
possibilities, but the comparisons are nonetheless instructive.
The procedures of Experiment 3 provided consistently stronger relations between delay and
left-key responding (r2 ranging from 0.93 to 0.99) than those of Experiments 1 and 2 (r2 ranging
from 0.48 to 0.92), but none of the fits was substantially or consistently superior to any of the
others. The respective mean r2 values for the exponential, hyperbolic and logarithmic functions
across all pigeons and experiments were, respectively, 0.804, 0.799 and 0.802, and these
differences would not even have been evident if the means had been rounded to only two
decimal places. Certainly the differences are well within the range of error to be expected from
behavioral preparations.
These findings seem to lead to the unsettling conclusion that we cannot resolve the issue of
which functions best characterize delay gradients by appealing to experimental data. It may be
useful to examine which functions would be most compatible with the fits to right-key data and
with those cases in which total response rates (left plus right) approximated constancy. Not to
keep our readers in suspense, we must state here that this issue will still be unresolved by the
time we conclude this report.
<Insert Table 2 about here>
Catania et al., Page 26
Experiment 4: Varying Delay by Varying Both Number and Time
One reason that these experiments began with number rather than time as the dimension
to fill the delay between left-key responding and the later reinforced right-key peck is that
responding rather than time is often implicated in important behavioral phenomena. The human
verbal learning literature, for example, is filled with many relevant cases, as when forgetting is
determined more strongly by intervening activities such as items read or problems solved than by
elapsed time (e.g., see Catania, 2013, Chapters 21, 26 and 27; White, 2001; Wright et al., 1990).
In addition, differential reinforcement can be effectively and independently applied to either
duration or number (Catania, 1991, pp. 13-17; LeFever, 1973).
Experiment 4 attempted to tease apart the contributions of number and time to delay. Its
rationale is illustrated in Figure 25. The left side of the figure shows what might happen to a
sequence of four lefts and four rights in the topographical tagging procedure if delay effects
depended solely on time, and the right side shows what might happen if delay effects depended
solely on number. The question is whether, given appropriate contingencies, the data conform to
the kinds of effects shown on the left or to those shown on the right. The example assumes
negligible induction (response generalization) of response rates from one key to the other.
<Insert Figure 25 about here>
At start, with no added contingencies (a and b), the performances are identical with
regard to both the time and the number of responses separating the last left from the later
reinforced right response, so here time-based delays and number-based delays are necessarily
equal. In c and d, a short-IRT contingency has shaped more rapid responding during the right
run; in c, these shorter IRTs mean a shorter time-based delay, which engenders faster left
Catania et al., Page 27
responding; but in d, where delay effects depend on number rather than time, the number-based
delay is unchanged, so the faster right responding has no effect on left responding. In e and f, the
right sequence is lengthened so that the run now takes as long as it had in a and b. Again, in e
time rather than number matters, so the time-based delay is equivalent to that in a and left
responding returns to its earlier a level; but in f the increased responding creates a longer
number-based delay, so left responding slows down. The inverse cases are shown in g through j.
With a long-IRT contingency in g, the time-based delay increases with slower right responding
and engenders slower left responding; but when the required right sequence is shortened in i so
that the time-based delay is the same as in a, left responding returns to its a level. In h, however,
the long-IRT contingency has no effect on the number-based delay, so left responding is
unaffected; but when in j the sequence is shortened, left responding increases because the
number-based delay has been reduced.
The target of this experiment, then, was to arrange contingencies for right-key runs that
would allow us to tell whether delays work like the time-based ones on the left side of the figure
or the number-based ones on the right side.
Method
Four White Carneaux in the UMBC laboratory, male Pigeons 60, 75, 77 and female Pigeon
81, served in daily sessions in a two-key chamber with the left key lit amber and the right key lit
green. As in prior experiments, the procedures were higher-order schedules in which RI 60-s
contingencies set up reinforcers for tandem left-key right-key sequences. This time, the left-key
requirement of the tandem schedules (M) was held constant at eight while right-key
contingencies varied. These variations, produced by the shaping of longer or shorter right-key
runs and of higher or lower right-key response rates within runs, were introduced in order
Catania et al., Page 28
independently to vary time-based delays and number-based delays as measured from the last left
peck to the reinforced right peck.
The right-key contingencies included, separately and in combination, variations in the
required number of right-key pecks (N), differential reinforcement of high right-key rates (drh),
in which the sum of the right-key interresponse times (IRTs) within a run had to be shorter than
some criterion value to be eligible for reinforcement, and differential reinforcement of low rightkey rates (drl), in which the final IRT of the right-key run had to be longer than some criterion
value to be eligible for reinforcement. Thus, if the tandem component consisted of FR 8 for the
left key (M=8) followed by a drh 2-s contingency imposed on FR 5 for the right key (N=5), a set
up reinforcer would be delivered given eight successive pecks on the left key followed by five
successive pecks on the left key emitted in less than 2 s (i.e., four IRTs averaging less than 0.5 s).
Similarly, if the tandem component consisted of FR 8 for the left key (M=8) followed by a drl
0.4-s contingency imposed on FR 4 for the right key (N=4), a set up reinforcer would be
delivered given eight successive pecks on the left key followed by four successive pecks on the
right key, the last two of which were separated by more than 0.4 s (i.e., IRT greater than 0.4 s).
The joint continua of time-based and number-based delays were explored for each pigeon.
At various points during shaping, contingencies were maintained over several sessions for
purposes of data collection. In most instances, data were collected only when performance was
judged stable according to the criteria of zero or one uncollected setups over successive sessions
and of visual inspection of the data, but the criteria were relaxed at extreme parameter values,
where it was sometimes difficult to maintain performance. In some instances, especially with drh
requirements, it was necessary to relax contingencies by dropping back to earlier parameter
values.
Catania et al., Page 29
Given these considerations and given that these shaping procedures were conducted over
more than two years of daily sessions for Pigeons 60, 75 and 77 and almost a year of daily
sessions for Pigeon 81, it would be impractical to present the details of each pigeon’s sessions
and data. Appendix 3 shows selected measures for each pigeon from the sessions that provided
the data figures to follow. For Pigeons 60, 75, 77 and 81 over the course of shaping sessions, the
respective ranges of right-key requirements were 1 to 12, 1 to 8, 1 to 12, and 1 to 16, the ranges
of drh requirements (sum of right IRTs) were 0.12 to 1.12 s, 0.16 to 1.40 s, 0.16 to 1.00 s, and
0.48 to 2.10 s, and the ranges of drl requirements (last right IRT) were 0.14 to 0.79 s, 0.12 to
0.77 s, 0.13 to 0.79 s, and 0.19 to 0.28 s.
After these procedures, all pigeons were given additional sessions of RI 60-s tandem FR 8
FR 1, which required only a single right response and therefore a delay equal to the most recent
left to right changeover (25, 8, 15 and 11 sessions respectively for each pigeon) and of an RI 60-s
schedule arranged for left-key responding only (15, 41, 24 and 21 sessions respectively for each
pigeon). The respective pairs of left response rates for the two condition were 58.5 / 55.3, 26.0 /
26.1, 45.0 / 33.2, and 50.4 / 48.5 responses per min. Except for the third pigeon, 77, the
differences between these two-key and one-key rates are small. We will not use these rates
elsewhere but provide them as possibly useful baselines for those who may wish to have them in
examining other data from these pigeons.
Results
Figure 26 shows data in XYZ coordinates from Pigeon 60 over the course of its shaping
procedures, where X equals the number of required right-key responses (N) between the last left
peck and the reinforcer (number delay), Y equals the time from the last left peck to the next
reinforced right peck (time delay), and Z equals the left-key response rate obtained at that
Catania et al., Page 30
intersection of X and Y or, in other words, at a given number-delay time-delay combination.
Data from sessions with no differential reinforcement of local right-key rates are shown as filled
circles; those from sessions with differential reinforcement of those local rates are shown as open
triangles with apex pointing up for drh sessions and down for drl sessions. The middle
coordinates are shown from a perspective that allows an overview of the relations among the
variables. The coordinates on the left are rotated so that they are seen from the number side,
thereby emphasizing the relation between response rates and number delay, whereas those on the
right are rotated so that the same data are seen from the time side, thereby emphasizing the
relation between response rates and time delay. From these perspectives, the number-based data
(left) show relatively shallow rate changes with changes in number whereas the time-based data
(right) provide a gradient similar to those obtained in the earlier experiments.
<Insert Figure 26 about here>
Figure 27 shows the data for each pigeon in XY coordinates, with left-key response rates
plotted against number of responses to the reinforcer (number delay) in the left column and time
to the reinforcer (time delay) in the right column. Sessions obtained with drh contingencies (apex
up triangles) drl contingencies (apex down triangles) or no rate contingencies (circles) arranged
for responding during right-key runs have been separately fitted with exponential functions (no
fit can be shown for response delay with the drl contingency for Pigeon 81 because all sessions
with that condition were conducted at a single N value, N = 4).
<Insert Figure 27 about here>
These graphs may be thought of as data projections against the rear wall of the XYZ
coordinates first when facing from the number side (left column) and then when facing from the
Catania et al., Page 31
time side (right column), and therefore corresponding to the perspective views shown in the left
and right sections of the triptych in Figure 26.
Although the data from Pigeon 75 are less orderly than those from the other pigeons, as a set
all support the argument that time delays more tightly control left-key responding than do
response delays. On the whole, the fits are closer together in the time-delay graphs on the right
than in the response-delay graphs on the left, and most of the fits on the left are of shallower
slope than those on the right, with one of the latter (Pigeon 75) even showing an increasing rather
than a decreasing function.
Discussion
The relation between response delays and response rates (left column) seems weaker than
that between time delays and response rates (right column), but it can still be argued from these
data that intervening response number, like intervening time, makes some contribution to the
effects of delayed reinforcers. This argument can be made even in the light of data such as those
of Dews’ interrupted fixed intervals (Figure 1), because at any time in the interval responding is
separated from the reinforcer by the intervening FI responses as well as by the remaining FI
time.
<Insert Figure 28 about here>
Decreased responding with increased number delays, however, are not sufficient evidence
that number is an effective delay variable. Figure 28 shows why this may be so. In the nine XYZ
coordinates, each row provides three views of a set of delay functions as seen from different
perspectives. In the top row, set A, Z decreases only with X and remains constant over changes
in Y (left column); a gradient is not evident with viewing from the Y side (middle column) but
one appears with viewing from the X side (right column). In the middle row, set B, Z decreases
Catania et al., Page 32
only with Y and remains constant over changes in X (left column); this time a gradient is evident
with viewing from the Y side (middle column) but not with viewing from the X side (right
column). In the bottom row, set C, Z is a joint function of X and Y, so viewing from either the Y
side or the X side shows decreasing Z with increases in the value of either the X or the Y
parameter.
Now consider the function defined by the filled circles, which are identical in the three XYZ
sets. If data samples of Z were obtained only along the diagonal of the XY base or, in other
words, only when X and Y were equal, then the outcome would not discriminate among the A, B
and C alternatives because each would generate identical data, as illustrated by the filled circles
in the various perspective views. Figure 29 provides an approximation to such a case in an XYZ
plot of the data of Pigeon 81. Over the course of shaping, number delays and time delays tended
to be correlated; this correlation across the data from all four pigeons is illustrated in Figure 30.
<Insert Figures 29 and 30 about here>
The implication is that flat functions should not have been expected in the left graph of
Figure 27. For example, delays tend to grow with increased numbers even with added drh
contingencies, so it is difficult to move behavior along a shaping trajectory that remains parallel
either to X or to Y in the XYZ coordinates. Given such constraints, the path of the XY
coordinates will to some extent move diagonally across the XY space during shaping. Responsedelay gradients that are not flat may be totally compatible with behavior that is determined solely
by time delay. The inverse, of course, is also a possibility.
One feature of the performances of our pigeons that is not evident in the summary data of
Appendix 3 is that it often became very difficult to budge responding once shaping had moved it
some way along either the time or the number continuum. Perhaps this was because at those
Catania et al., Page 33
points we were approaching limits on the behavior that can be supported by a given delay of
reinforcement. One way in which that could happen has been demonstrated within a simulation
of operant behavior built upon some of the assumptions about delay that we have been exploring
here (cf. Catania, 2005e, Section 3.2).
If delay gradients involved well-defined limits on behavior, then variability is a relevant
dimension in judging the priorities of accounts in terms of response delay or in terms of time
delay. Table 3 shows r2 values for the functions shown fitted to the data in Figure 27. Based on
different assumptions about the distributions underlying exponential functions, more than one
way to calculate their variances accounted for is available (e.g., Cameron & Windmeijer, 1997);
those presented here were generated by curve-fitting programs in DeltaGraph7© (Red Rock
Software).
<Insert Table 3 about here>
In ten of the eleven possible comparisons available in Table 3, r2 was larger for delays
measured in terms of time than for delays measured in terms of number, sometimes by large
magnitudes (e.g., Pigeon 60, condition Lo; Pigeon 75, condition Hi; and Pigeon 81, condition O).
A difference in the same direction is also evident in comparing the data from the number-based
procedures of Experiments 1 and 2 (Figures 13 and 21) with those of the time-based procedure of
Experiment 3 (Figure 23). The weight of the evidence implies that time, not number, is the
appropriate metric for assessing delay of reinforcement.
Experiment 5: Delay and Reinforcer Duration
The research up to this point has been conducted with fixed reinforcer durations. The
literature suggests that reinforcer magnitude, and duration in particular, is not typically a
Catania et al., Page 34
powerful variable, especially when compared with rate of reinforcement (e.g., Catania, 1963;
Doughty, Galuska, Dawson, & Brierley, 2012; Grace, 1995; Jenkins & Clayton, 1949; Ludvig,
Conover, & Shizgal, 2007). Some properties of delay gradients that have been considered here
suggest why this may be the case.
Consider a thought experiment in which a fixed-interval (FI) schedule is arranged for
responding maintained by a 3-s reinforcer. Hypothetical gradients that might be in play in the
procedure are shown in Figure 31. First, the schedule operates by itself (A) and engenders a
characteristic gradient. Then a second schedule is introduced that provides another 3-s reinforcer
partway through the first interval (B). That reinforcer truncates the tail of the original gradient
(cf. Figure 9), but responding is given by the total area under the two segments, so the two 3-sec
reinforcers together generate more responding than had been maintained by the single 3-s
reinforcer alone in A. Next, the added reinforcer is moved closer to the end of the original
interval (C) and then closer still (D), each time cutting off a greater area of the original gradient
than the tail of the new gradient brings in. Eventually, the interpolated reinforcer is scheduled so
late in the original interval that the second 3-s reinforcer follows almost immediately after the
earlier one (E). At this point, the two have effectively become a single 6-s reinforcer, but the
gradient maintained by the original 3-s reinforcer has been almost totally truncated. Because
maintained responding is given by the total area under the gradient, this new 6-s reinforcer
maintains about the same quantity of behavior as had the 3-s reinforcer when it was all by itself.
<Insert Figure 31 about here>
If delay gradients operate consistently with the gradients of this thought experiment, then
changing reinforcer durations for the performances we have been studying should have little if
any effect on response rates. Experiments in which changes in the value of some parameter have
Catania et al., Page 35
no effect are ordinarily of little interest, but in this instance it is worthwhile to confirm or
disconfirm the predicted outcome.
Method
Three White Carneaux in the UMBC laboratory, Pigeons 73,, 74 and 76, served in daily
sessions in a two-key pigeon chamber. As in Experiments 2 and 3, their pecking was maintained
by the RI 60-s reinforcement of a tandem FR M (left key) FR N (right key) sequence; M was
held constant at four and N at eight. The maximum session duration was 60 min, but maximum
reinforcers per session was adjusted so as to reduce the differences in postsession feeding that
would otherwise have followed from the different reinforcer durations. Successive conditions for
each pigeon are outlined in Table 4, which also summarizes the data.
<Insert Table 4 about here>
Given that response rates sometimes change systematically within sessions (McSweeney &
Hinson, 1992; McSweeney, Roll, & Cannon, 1994; Weatherly, McSweeney, & Swindell, 1995),
response rates calculated over the entire session were supplemented with rates calculated over
only the first twenty min of each session, but in each case while still adhering to the constraints
on data collection illustrated in Figure 11. Stability criteria and other details were as described in
General Method.
Results
In Figure 32, left-key response rates are shown for the first twenty minutes of each session
and for the entire session, averaged over the last five sessions of each condition for each pigeon.
Response rates measured over the first twenty min were higher than rates measured over the full
session, indicated that response rates early in the session tended to drop off as the session
continued. Despite these differences in magnitude, however, the relations between response rates
Catania et al., Page 36
and reinforcer durations were similar across the two measures: left-key responding did not vary
consistently with reinforcer duration. Linear fits showed positive slopes for Pigeon 73 and
negative slopes for Pigeons 74 and 76,
<Insert Figure 32 about here>
Discussion
Usually a finding of no reliable differences across changes in the value of some parameter is
of minimal interest. In this case, however, the finding is consistent both with prior data and with
the implications of our assumptions about how delayed reinforcement operates. In particular, if
an interpolated reinforcer engenders its own gradient and blocks the gradients engendered by
later reinforcers from acting upon the responses that preceded it, as in Figures 9 and 21, then it
follows that the onset of a reinforcer rather than its duration may be the major source of its
potency. If this is how it works, it should be no surprise that small reinforcers can have large
effects.
Experiment 6: Delay and the Random-Interval Parameter
The delay gradients in the research so far have all been obtained with an RI 60-s schedule
maintaining the higher-order terminal sequence of responses on the left and right keys. Longer or
shorter RI schedules would not only provide lower or higher overall rates of reinforcement; they
would also allow longer or shorter tails in the delay gradient, and perhaps they would also have
other effects. We have assumed that beyond the limits of the left-key runs specified in our
contingencies, gradients will be shallow and also that, at that remove and therefore not being
subject to those contingencies, responses will consist of a mix of responses on the two keys,
probably approximating the mix observed in the relative rates of responding on the two keys.
Catania et al., Page 37
The implication is that changes in RI value should have small effects, if any, on the delayu
gradients obtained with topographical tagging procedures. Once again, as with Experiment 5, it
may be of interest to examine a parameter changes in the value of which are expected to have
little or no effect. This last experiment explored the RI parameter both by varying delays with RI
15-s, RI 30-S and RI 60-s schedules, and by holding delay roughly constant with a fixed M plus
N sequence while the RI value was varied over a range from 5 to 120 s.
Method
Four White Carneaux, male Pigeons 60, 73, 76 and female Pigeon 77, and two Silver King
males, Pigeons 47 and 59, served in daily sessions in two-key chambers in the UMBC
laboratory. Procedures were similar to those of earlier experiments. In the first, with the required
left-key requirement (M) held constant at 4, right-key requirements (N) of 2, 4 and 8 were
examined during sessions of RI t-s (tandem FR M FR N) schedules. Pigeons 73 and 77 served in
one two-key chamber, with t set to 15, 30 and 60 s; the maximum session duration was set to 30
min across all condition for these two pigeons, to reduce the impact of changes in response rate
within sessions (cf. Experiment 5, Method). Pigeons 47 and 59 served in a different two-key
chamber in sessions at a different time, with t set either to 15 s or to 60 s, and with maximum
session duration set to 30 min for RI 15-s and 75 min for RI 60-s. In all conditions for all
pigeons, maximum reinforcers per session was set to 60. Order of conditions and other details
are provided in Table 5.
<Insert Table 5 about here>
In a second procedure, conducted in still another two-key chamber with Pigeons 60, 73 and
76, the respective left-key and right-key requirements of the tandem component of the schedule
Catania et al., Page 38
were held constant at M = 8 and N = 4, and the RI t-s schedule was varied over a range of t from
5 to 120. Table 6 provides order of conditions and other details.
<Insert Table 6 about here>
Results
Figure 33 shows left-key response rates as a function of the delay from the last left peck to
the next reinforced right peck, with the value of the RI schedule as a parameter. Exponential
functions have been fitted to each data set. No systematic differences in the slopes of the
functions, in their ordering along the y-axis or in variability are evident across the four pigeons.
<Insert Figure 33 about here>
The left column of Figure 34 shows left-key and right-key response rates as a function of the
value of the random interval, with the tandem schedule held constant at M=8 and N=4. The
functions decrease with RI value for Pigeon 60, increase for Pigeon 73, and also increase, though
only slight, for Pigeon 76. The right column of the figure shows that the changes in rate were
correlated with corresponding inverse changes in the delays to reinforcement with changes in RI
value. Not only were the effects of changes in the RI parameter unsystematic across pigeons, but
those changes that did occur are probably better attributed to varying delays of reinforcement
over conditions than to any direct effects of RI value per se.
<Insert Figure 34 about here>
Discussion
Once again we have a finding of no reliable differences across changes in the value of a
parameter. In this instance the finding is reassuring, because it implies that data-based
generalizations about delay gradients engendered by RI scheduling need not be limited to a
Catania et al., Page 39
narrow range of schedule values. Whether a similar generality will hold with respect to other
parameters remains to be seen.
General Discussion
Our findings with topographical tagging have direct implications for the analysis of
performances maintained by concurrent schedules of reinforcement. Accounts in terms of the
changing reinforcement probabilities for continued responding on one key and for changing over
to the other (Shimp, 1966) need to be expanded, because each reinforcer operates on prior
responding on both keys, no matter which peck produced it. A concern about such extended
effects of reinforcers was in fact implicit in supplements to concurrent procedures such as the
changeover delay (COD), which prevented a response on one key from being followed
immediately by a reinforced response on the other key (e.g., Catania, 1976). The interactions
among concurrent responses and their reinforcers provide a challenge for computer simulations
(Catania, 2005e); those that treat reinforcement and shaping as a variety of selection seem
especially promising (McDowell, 2014).
If delay effects depend mainly on the onset of reinforcers, as implied by the results of
Experiment 5, then even small reinforcers can be potent agents of behavior change. One such
event is the onset of a discriminative stimulus, and one response that may be reinforced by a
discriminative stimulus is looking at it. In observing-response procedures (Dinsmoor, 1983,
1989; Kelleher, Riddle, & Cook, 1962), seeing the stimulus is not an appreciably delayed
consequence but the reinforcers later produced in its presence are, and the durations of their
delays may determine the potency of the stimulus as a reinforcer of looking or observing
responses (Grice, 1948).
Catania et al., Page 40
Figure 35 shows that delay gradients maintained by discriminative stimuli as reinforcers are
similar in form to those maintained by consumable reinforcers, such as those examined here. The
left key in a two-key chamber was an observing-response key; the right key was a schedule key
(Catania, 2012). Equal proportions of FI or extinction (EXT) components operated during yellow
on the right key, unless observing responses, pecks during a pre-schedule stimulus, changed the
contingencies so that, instead of yellow for both, green accompanied the FI schedule and red
accompanied EXT. In this procedure, therefore, changes in the FI value changed the delay
between the onset of green on the right key and the later delivery of the FI reinforcer.
<Insert Figure 35 about here>
Delay gradients engendered by a range of reinforcers, including stimulus onset in the
maintenance of observing behavior, have been incorporated into a model of ADHD (attentiondeficit hyperactivity disorder) by Sagvolden and colleagues (Catania, 2005a, 2005b; Sagvolden,
Johansen, Aase, & Russell, 2005). In that account, delay gradients that decrease more rapidly
than those in a general population can lead to hyperactivity when they differentially reinforce
rapid response sequences, because the gradients cannot support the earlier responses of those
sequences if responding is emitted more slowly. They can also lead to attention deficit when
those gradients reduce the likelihood of observing responses, because stimulus onset becomes a
less effective reinforcer if responding does not quickly produce reinforcers in the presence of that
stimulus. Furthermore, consistent with ADHD as studied in populations of children, the model
shows that different delay-gradient slopes can engender different proportions of the hyperactivity
and attention-deficit components of the syndrome.
A frequent component of experimental procedures that involve complex discriminations is a
correction procedure, which prevents a trial containing errors from being promptly followed by a
Catania et al., Page 41
trial containing a reinforced correct response. The recognition that reinforcers produced by
correct responses can also act on earlier incorrect responses is implicit in this procedure.
Similar contingencies operate in educational environments, so applied practitioners and
teachers in general must be alert for situations in which they might be strengthening incorrect
responses along with the corrects that they reinforce (cf. Catania, 2011a). Reinforcing every
correct response and repeating any trial with errors until a child gets it right guarantees that any
error sequence will eventually be followed by a reinforced correct response. The correct
responses will probably dominate eventually, because they are most closely followed by
reinforcers. But errors might diminish slowly and perhaps continue indefinitely at modest levels
because they are reliably followed later by reinforced correct responses.
When errors or problem behavior share in the effects of a reinforcer, we may mistakenly
conclude that the reinforcer is not doing its job very well. But if the reinforced behavior includes
response classes that we did not intend to reinforce, it may simply be doing very well a job other
than the one we wanted it to do. Behavior that we do not want to reinforce should be kept from
getting consistently close to reinforcers produced by other responses.
Delay of reinforcement undoubtedly has many more implications for and applications to
human behavior (Stromer, McComas, & Rehfeldt, 2000). Another example of course involves
the effects of intersecting steeper and shallower delay gradients, leading to analogues of the
behavior that we discuss in terms of self-control (Rachlin & Green, 1972). But any extensions to
delay discounting with humans (e.g., Green & Myerson, 2004; Johnson & Bickel, 2002) are most
likely tenuous and indirect. Somehow the delay gradients that interact with behavior in
nonhuman preparations must be converted to mathematical or other nontemporal relations when
humans deal with temporally distant events and discount remote consequences more than those
Catania et al., Page 42
more imminent. It seems unlikely that this problem can be addressed without an account of
verbal behavior, and that may be more attainable with research on how relevant histories of
mathematical behavior are created than by studying relations obtained from populations for
which the relevant history is probably variable and mostly inaccessible. But the topic is crucial,
because verbal behavior seems to allow us to bypass some of the otherwise inevitable effects of
delayed reinforcers.
Delay of reinforcement is ubiquitous, and these experiments have only scratched the surface
of issues related to it. For example, we do not know enough about whether its effect depends on
how responses that occur during the delay are related to the response that produces the
reinforcer, as in the concept of coupling (Killeen, 1994; Killeen & Sitomer, 2003). The effect
produced by the relation between the reinforcer and the response that produced it seems to have
special potency. Does that imply that a response like the one that produced the reinforcer will be
strengthened more than one that is very different? Might the creation of one operant class
insulate members of that class from the delayed effects of reinforcers later produced by members
of another class?
We also do not know enough about how delayed reinforcement might be involved in the
knitting together of temporally extended response sequences, so they may become behavioral
units in their own right (e.g., the distinction between chaining and chunking; see Catania, 2013,
pp. 127-129). We do not know enough about the effects of delayed reinforcement on variability
(Neuringer, 2002; Odum, Ward, Barnes, & Burke, 2006; Wagner & Neuringer, 2006). Does
responding become more or less variable with increasing temporal separation from the
reinforcer? We do not know enough about the role of delayed reinforcers in transitions from
schedules in which reinforcers are contingent upon responses to those of free or adventitious
Catania et al., Page 43
reinforcer deliveries (Catania, 2005d; Sizemore & Lattal, 1977; Williams & Lattal, 1999). We do
not know enough about tradeoffs between delayed reinforcement and changes in reinforcement
probability (Aucella, 1984; Lattal, 1974). And we do not know enough about the effects of
delayed reinforcers on measures other than response rate, such as accuracy (Blough, 1959),
choice (Fantino & Abarca, 1985; Gentry & Marr, 1980; Hackenberg, 1992; Mazur, 1995), or
resistance to change (Doughty & Lattal, 2003; Nevin & Grace, 2000; Podlesnik & Shahan,
2008). We could multiply such questions, but one thing we do know: More research on delay of
reinforcement is needed.
Acknowledgments
Many have contributed in many different ways to the research reported here. We especially
thank Rosie Mills, the UMBC Animal Caretaker. A number of UMBC undergraduates over
several acsdemic years assisted with a variety of tasks related to daily experimental sessions,
animal care, and the maintenance of data and records. These included, among others, Pearl
Friedman, Megan Glaze, Mariko Johnson, Stacey Long, Kevin Stine, Irina Vishnevetsky, and
Crescianciana Malinis. Experiment 5 was derived from an undergraduate Honors project by Lara
Kowalsky Kehle, and Leanne Valentine contributed both to Experiments 1, 2 and 3 and to the
research reported in Figure 35. Koji Hori of Rikkyo University in Japan also participated during
a academic sabbatical at UMBC. The statistical and data analysis skills of Eric J. French of CMU
smoothed the way for the development of Table 2. We have undoubtedly omitted some others,
but we are grateful to them along with those we were able to list here.
Our departments at both campuses also deserve thanks. The UMBC department in particular
allowed us to keep working even through times with negligible outside funding of the research.
Catania et al., Page 44
As grants for operant research became increasingly hard to come by, one set of proposal reviews
declined support on the grounds that the experiments were not feasible, and that even if they
were feasible the results would be of little interest. Delays of reinforcement sometimes seemed to
be intrinsically embedded within the entire enterprise.
We must also thank our pigeons for their service. Among them, female Pigeon 77 and male,
Pigeon 81 had long been housed in adjacent cages. With a change in experimental procedures
one of them was moved so they were no longer next to each other. Both stopped working in their
experimental chambers. They only got back to work when we moved them next to each other
again. Upon the closing of the UMBC pigeon laboratory we were able to place all our pigeons
with a local pigeon fancier, who inquired about their histories. There is more to pigeon behavior
than pecking and eating. We were pleased to learn that he housed both of these pigeons together
in a roomy pigeon coop.
Last but certainly not least, we regret that Eliot Shimoff did not survive to contribute to or
even to see this product of our collaborative research. His laboratory skills, his programming
expertise, and above all his enthusiasm were invaluable. This research could not have been done
without him.
Catania et al., Page 45
References
Aucella, A. (1984). Delay versus probability of reinforcement in discrimination learning. Jouranl of
Experimental Psychology: Animal Behavior Processes, 10, 452-460.
Baum, W. M. (2004). Molar and molecular views of choice. Behavioural Processes, 66, 349-359.
Baum, W. M. (2012). Rethinking reinforcement: Allocation, induction, and contingency. Journal of the
Experimental Analysis of Behavior, 97, 101-124.
Berg, J. P., & McDowell, J. J. (2011). Quantitative, steady-state properties of Catania's computational
model of the operant reserve. Behavioural Processes, 75, 71-83.
Blough, D. S. (1959). Delayed matching in the pigeon. Journal of the Experimental Analysis of Behavior,
2, 151-160.
Cameron, A. C., & Windmeijer, F. A. G. (1997). An R-squared measure of goodness of fit for some
common nonlinear regression models. Journal of Econometrics, 77, 329-342.
Catania, A. C. (1963). Concurrent performances: A baseline for the study of reinforcement magnitude.
Journal of the Experimental Analysis of Behavior, 6, 299-300.
Catania, A. C. (1969). Concurrent performances: Inhibition of one response by reinforcement of another.
Journal of the Experimental Analysis of Behavior, 12, 731-744.
Catania, A. C. (1970). Reinforcement schedules and psychophysical judgments: A study of some
temporal properties of behavior. In W. N. Schoenfeld (Ed.), The theory of reinforcement schedules
(pp. 1-42). New York: Appleton-Century-Crofts.
Catania, A. C. (1971). Reinforcement schedules: The role of responses preceding the one that produces
the reinforcer. Journal of the Experimental Analysis of Behavior, 15, 271-287.
Catania, A. C. (1973). Self-inhibiting effects of reinforcement. Journal of the Experimental Analysis of
Behavior, 19, 517-526.
Catania, A. C. (1976). Concurrent performances: Rate constancies without changeover delays. Journal of
the Experimental Analysis of Behavior, 25, 377-387.
Catania, A. C. (1983). Is not-responding behavior? Behavioral and Brain Sciences, 6, 321-322.
Catania et al., Page 46
Catania, A. C. (1991). Time as a variable in behavior control. In I. H. Iversen & K. A. Lattal (Eds.),
Experimental analysis of behavior. Part 2 (pp. 1-19). Amsterdam: Elsevier.
Catania, A. C. (2002). The watershed years of 1958-1962 in the Harvard Pigeon Laboratory. Journal of
the Experimental Analysis of Behavior, 77, 327-345.
Catania, A. C. (2005a). Attention-deficit/hyperactivity disorder (ADHD): Delay-of-reinforcement
gradients and other behavioral mechanisms. Behavioral and Brain Sciences, 28, 419-424.
Catania, A. C. (2005b). Attention-deficit/hyperactivity disorder (ADHD): One process or many?
Behavioral and Brain Sciences, 28, 446-450.
Catania, A. C. (2005c). Eliot Shimoff, 1943-2004. Journal of the Experimental Analysis of Behavior, 83,
197-199.
Catania, A. C. (2005d). The nonmaintenance of behavior by noncontingent reinforcement. European
Journal of Behavior Analysis, 6, 89-94.
Catania, A. C. (2005e). The operant reserve: A computer simulation in (accelerated) real time.
Behavioural Processes, 69, 257-278.
Catania, A. C. (2011a). Basic operant contingencies: Main effects and side effects. In W. W. Fisher, C. C.
Piazza & H. S. Roane (Eds.), Handbook of applied behavior analysis (pp. 34-54). New York:
Guilford.
Catania, A. C. (2011b). Model versus model or model versus data? A commentary on Berg and
McDowell’s “Quantitative, steady-state properties of Catania’s computational model of the operant
reserve.”. Behavioural Processes, 87, 84-87.
Catania, A. C. (2012). The pursuit of experimental analysis. European Journal of Behavior Analysis, 13,
269-280.
Catania, A. C. (2013). Learning (5th ed.). Cornwall-on-Hudson, NY: Sloan.
Catania, A. C., de Souza, D., & Ono, K. (2004). Free-choice preference when one alternative is rarely or
never chosen. Braxilian Journal of Behavior Analysis, 1, 69-80.
Catania et al., Page 47
Catania, A. C., & Keller, K. J. (1981). Contingency, contiguity, correlation, and the concept of causation.
In P. Harzem & M. D. Zeiler (Eds.), Predictability, correlation, and contiguity. Advances in analysis
of behaviour, Volume 2 (pp. 125-167). Chichester: Wiley.
Catania, A. C., & Reynolds, G. S. (1968). A quantitative analysis of the behavior maintained by interval
schedules of reinforcement. Journal of the Experimental Analysis of Behavior, 11, 327-383.
Catania, A. C., Sagvolden, T., & Keller, K. J. (1988). Reinforcement schedules: retroactive and proactive
effects of reinforcers inserted into fixed-interval performances. Journal of the Experimental Analysis
of Behavior, 49, 49-73.
Chung, S.-H. (1965). Effects of delyed reinforcement in a concurrent situation. Journal of the
Experimental Analysis of Behavior, 8, 439-444.
Church, R. M., & Gibbon, J. (1982). Temporal generalization. Jouranl of Experimental Psychology:
Animal Behavior Processes, 8, 165-186.
Dews, P. B. (1960). Free-operant behavior under conditions of delayed reinforcement: I. CRF-type
schedules. Journal of the Experimental Analysis of Behavior, 3, 221-234.
Dews, P. B. (1962). The effect of multiple SΔ periods on responding on a fixed-interval schedule. Journal
of the Experimental Analysis of Behavior, 5, 369-374.
Dews, P. B. (1966). The effect of multiple SΔ periods on responding on a fixed-interval schedule: V.
Effect of periods of complete darkness and of occasional omissions of food presentation. Journal of
the Experimental Analysis of Behavior, 9, 573-578.
Dews, P. B. (1970). The theory of fixed-interval responding. In W. N. Schoenfeld (Ed.), The theory of
reinforcement schedules (pp. 43-61). New York: Appleton-Century-Crofts.
Dinsmoor, J. A. (1983). Observing and conditioned reinforcement. Behavioral and Brain Sciences, 6,
693-728.
Dinsmoor, J. A. (1989). Selective observing: Pigeons turn discriminative stimuli on and off by pecking
separate keys. In N. W. Bond & D. A. T. Siddle (Eds.), Psychobiology: Issues and Applications (pp.
443-456). North-Holland: Elsevier.
Catania et al., Page 48
Doughty, A. H., Galuska, C. M., Dawson, A. E., & Brierley, K. P. (2012). Effects of reinforcer magnitude
on response acquisition with unsignaled delayed reinforcement. Behavioural Processes, 90, 287-289.
Doughty, A. H., & Lattal, K. A. (2003). Response persistence under variable-time schedules following
immediate and unsignalled delayed reinforcement. Quarterly Journal of Experimental Psychology
Section B, 56, 267-277.
Fantino, E., & Abarca, N. (1985). Choice, optimal foraging, and the delay-reduction hypothesis.
Behavioral and Brain Sciences, 8, 315-330.
Ferster, C. B., & Skinner, B. F. (1957). Schedules of reinforcement. New York: Appleton-Century-Crofts.
Gentry, G. D., & Marr, M. J. (1980). Choice and reinforcement delay. Journal of the Experimental
Analysis of Behavior, 33, 27-37.
Grace, R. C. (1995). Independence of reinforcement delay and magnitude in concurrent chains. Journal of
the Experimental Analysis of Behavior, 63, 255-276.
Green, L., & Myerson, J. (2004). A discounting framework for choice with delayed and probabilistic
rewards. Psychological Bulletin, 130, 769-792.
Grice, G. R. (1942). An experimental study of the gradient of reinforcement in maze learning. Journal of
Experimental Psychology, 30, 475-489. doi: 10.1037/h0055908
Grice, G. R. (1948). The relation of secondary reinforcement to delayed reward in visual discrimination
learning. Journal of Experimental Psychology, 38, 1-16. doi: 10.1037/h0061016
Hackenberg, T. D. (1992). Choice in situations of time-based diminishing returns: Immediate versus
delayed consequences of action. Journal of the Experimental Analysis of Behavior, 57, 67-80.
Hamilton, E. L. (1929). The effect of delayed incentive on the hunger drive in the white rat. Genetic
Psychology Monographs.
Hineline, P. N. (2001). Beyond the molar-molecular distinction: We need multiscaled analyses. Journal of
the Experimental Analysis of Behavior, 75, 342-347.
Hull, C. L. (1943). Principles of behavior. New York: Appleton-Century-Crofts.
Catania et al., Page 49
Jenkins, W. O., & Clayton, F. L. (1949). Rate of responding and amount of reinforcement. Journal of
Comparative and Physioogicval Psychology, 42, 174-181.
Johnson, M. W., & Bickel, W. K. (2002). Within-subject comparison of real and hypothetical money
rewards in delay discounting. Journal of the Experimental Analysis of Behavior, 77, 129-146.
Katz, I., & Catania, A. C. (2005). Concurrent performances: Extinction, noncontingent reinforcement and
variably delayed reinforcement. European Journal of Behavior Analysis, 6, 95-108.
Kelleher, R. T., Riddle, W. C., & Cook, L. (1962). Observing responses in pigeons. Journal of the
Experimental Analysis of Behavior, 5, 3-13.
Keller, F. S., & Schoenfeld, W. N. (1950). Principles of psychology. New York: Appleton-CenturyCrofts.
Killeen, P. R. (1988). The reflex reserve. Journal of the Experimental Analysis of Behavior, 50, 319-331.
Killeen, P. R. (1994). Mathematical principles of reinforcement. Behavioral and Brain Sciences, 17, 105172.
Killeen, P. R. (2001). Writing and overwriting short-term memory. Psychonomic Bulletin and Review,
8(1), 18-43.
Killeen, P. R. (2011). Models of trace decay, eligibility for reinforcement, and delay of reinforcement
gradients, from exponential to hyperboloid. Behavioural Proceeses, 87, 57-63.
Killeen, P. R., & Sitomer, M. T. (2003). MPR. Behavioural Processes, 62, 49-64.
Laties, V. G., Weiss, B., Clark, R. L., & Reynolds, M. D. (1965). Overt "mediating" behavior during
temporally spaced responding. Journal of the Experimental Analysis of Behavior, 8, 107-116.
Lattal, K. A. (1974). Combinations of response reinforcer dependence and independence. Journal of the
Experimental Analysis of Behavior, 22, 357-362.
Lattal, K. A. (1984). Signal functions in delayed reinforcement. Journal of the Experimental Analysis of
Behavior, 30, 11-17.
Lattal, K. A. (2010). Delayed reinforcement of operant behavior. Journal of the Experimental Analysis of
Behavior, 93, 129-139.
Catania et al., Page 50
LeFever, F. F. (1973). Instrumental response chains and timing behavior. (Doctoral Dissertation), New
York University, University Microfilms, Ann Arbor, MI. (73-19,943)
Ludvig, E. A., Conover, K., & Shizgal, P. (2007). The effects of reinforcer magnitude on timing in rats.
Journal of the Experimental Analysis of Behavior, 87, 201-218.
Mazur, J. E. (1987). An adjusting procedure for studying delayed reinforcement. In M. L. Commons, J. E.
Mazur, J. A. Nevin & H. Rachlin (Eds.), Quantitative Analyses of Behavior, Vol. 5, The Effect of
Delay and of Intervening Events on Reinforcement Value (pp. 283-308). Hillsdale, NJ: Erlbaum.
Mazur, J. E. (1995). Conditioned reinforcement and choice with delayed and uncertion primary
reinforcers. Journal of the Experimtal Analysis of Behaivor, 63, 139-150.
McDowell, J. (2014). Predictions of an evolutrionary theory of behavior dynamics. Paper presented at the
Society for the Quantitative Analyses of Behavior, Chhicago, IL.
McSweeney, F. K., & Hinson, J. M. (1992). Patterns of responding within sessions. Journal of the
Experimental Analysis of Behavior, 58, 19-36.
McSweeney, F. K., Roll, J. M., & Cannon, C. B. (1994). The generality of within-session patterns of
responding: Rate of reinforcement and session length. Animal Learning and Behavior, 22, 252-266.
Murre, J. M. J., & Chessa, A. G. (2011). Power laws from individual differences in learning and
forgetting: Mathematical analyses. Psychonomic Bulletin and Review, 18, 592-597.
Neuringer, A. J. (1969). Delayed reinforcement versus reinforcement after a fixed interval. Journal of the
Experimental Analysis of Behavior, 12, 375-383.
Neuringer, A. J. (2002). Operant variability: Evidence, functions, and theory. Psychonomic Bulletin and
Review, 9, 672-705.
Nevin, J. A., & Grace, R. C. (2000). Behavioral momentum and the Law of Effect. Behavioral and Brain
Sciences, 23, 73-130.
Odum, A. L., Ward, R. D., Barnes, C. A., & Burke, K. A. (2006). The effects of delayed reinforcement on
variability and repetition of response sequences. Journal of the Experimental Analysis of Behavior,
86, 159-179.
Catania et al., Page 51
Pavlov, I. P. (1927). Conditioned reflexes (G. V. Anrep, Trans.). London: Oxford University Press.
Perin, C. T. (1943). A quantitative investigation of the delay-of-reinforcement gradient. Journal of
Experimental Psychology, 32, 37-51.
Podlesnik, C. A., Jimenez-Gomez, C., Ward, R. D., & Shahan, T. A. (2006). Resistance to change of
responding maintained by unsignaled delays to reinforcement: A response-bout analysis. Journal of
the Experimental Analysis of Behavior, 85, 329-347.
Podlesnik, C. A., & Shahan, T. A. (2008). Response–reinforcer relations and resistance to change.
Behavioural Processes, 77, 109-125.
Rachlin, H., & Green, L. (1972). Commitment, choice and self-control. Journal of the Experimental
Analysis of Behavior, 17, 15-22.
Reilly, M. P., & Lattal, K. A. (2004). Within-session delay-of-reinforcement gradients. Journal of the
Experimental Analysis of Behavior, 82, 21-35.
Renner, K. E. (1964). Delay of reinforcement: A historical review. Psychological Bulletin, 61, 341-361.
doi: 10.1037/h0048335
Sagvolden, T., Johansen, E. B., Aase, H., & Russell, V. A. (2005). A dynamic developmental theory of
Attention-Deficit/Hyperactivity Disorder (ADHD) predominantly hyperactive/impulsive and
combined subtypes. Behavioral and Brain Sciences, 28, 397-468.
Seward, J. P. (1942). The delay-of-reinforcement gradient in maze learning. Journal of Experimental
Psychology, 30, 464-474.
Shahan, T. A., & Lattal, K. A. (2005). Unsignaled delay of reinforcement, relative time, and resistance to
change. Journal of the Experimental Analysis of Behavior, 83, 201-219.
Shimp, C. (1966). Probabilistically reinforced choice behavior in pigeons. Journal of the Experimental
Analysis of Behavior, 9, 443-455.
Sizemore, O. J., & Lattal, K. A. (1977). Dependency, temporal contiguity, and response-independent
reinforcement. Journal of the Experimental Analysis of Behavior, 27, 119-125.
Catania et al., Page 52
Skinner, B. F. (1938). The behavior of organisms: An experimental analysis. New York: AppletonCentury-Crofts.
Skinner, B. F. (1940). The nature of the operant reserve. Psychological Bulletin, 37, 423.
Staddon, J. E. R. (1965). Some properties of spaced responding in pigeons. Journal of the Experimental
Analysis of Behavior, 8, 19-27.
Staddon, J. E. R., & Simmelhag, V. L. (1971). The "superstition" experiment: A reexamination of its
implications for the principles of adaptive behavior. Psychological Review, 78, 3-43.
Steinman, W. M. (1968). Response rate and varied reinforcement: Reinforcers of similar strengths.
Psychonomic Science, 10, 35-36.
Stromer, R., McComas, J., & Rehfeldt, R. A. (2000). Designing interventions that include delayed
reinforcement: Implications of recent laboratory research. Journal of Applied Behavior Analysis, 33,
359-371.
Wagner, K., & Neuringer, A. (2006). Operant variability when reinforcement is delayed. Learning and
Behavior, 34, 111-123.
Watson, J. B. (1917). The effect of delayed feeding upon learning. Psychobiology, 1, 51-59.
Weatherly, J. N., McSweeney, F. K., & Swindell, S. (1995). On the contributions of responding and
reinforcement to within-session patterns of responding. Learning and Motivation, 421-432.
White, K. G. (2001). Forgetting functions. Animal Learning and Behavior, 29, 193-207.
Williams, A. M., & Lattal, K. A. (1999). The role of the response-reinforcer relation in delay-ofreinforcement effects. Journal of the Experimental Analysis of Behavior, 71, 187-194.
Wixted, J. T., & Gaitan, S. C. (2002). Cognitive theories as reinforcement history surrogates: The case of
likelihood ratio models of human recognition memory. Animal Learning and Behavior, 30, 289-305.
Wolfe, J. B. (1934). The effect of delayed reward upon learning in the white rat. Journal of Comparative
Psychology, 17, 1-21.
Catania et al., Page 53
Wright, A. A., Cook, R. G., Rivera, J. J., Shyan, M. R., Neiworth, J. J., & Jitsumori, M. (1990). Naming,
rehearsal, and interstimulus interval effects in memory processing. Journal of Experimental
Psychology: Learning, Memory, and Cognition, 16, 1043-1059.
Catania et al., Page 54
Authors Note
The first author may be contacted at catania@umbc.edu and the second author at
reill1mp@cmich.edu.
<No footnotes>
Catania et al., Page 55
Table 1. Left-key (L) and right-key (R) responding in Experiment 3.
Data for three pigeons. PRP = post-reinforcement pause, rfr = reinforcer. Details in text.
Delay
Pigeon F! (s) (s)
11 0.25
0.99
0.50
1.50
1.00
2.55
2.00
2.92
4.00
5.15
L
resp/min
73.5
51.7
28.9
30.8
12.1
R
resp/min
14.5
9.1
5.5
13.7
20.3
L+R
resp/min
88.0
60.8
34.4
44.5
32.4
PRP
rfr/min (s)
Sessions
1.09
3.15
25
1.01
4.24
56
1.07
5.71
37
1.07
7.43
21
0.99
11.03
33
Delay
F! (s) (s)
54 0.25
0.59
0.50
0.87
1.00
1.14
2.00
2.16
4.00
4.34
L
resp/min
67.4
52.9
45.5
29.2
12.5
R
resp/min
15.5
17.0
34.0
61.6
35.5
L+R
resp/min
82.9
69.9
79.5
90.8
48.0
PRP
rfr/min (s)
Sessions
1.06
2.18
22
0.96
2.39
21
1.10
3.26
20
1.02
6.34
21
0.85
11.75
38
Delay
F! (s) (s)
64 0.25
0.83
0.50
0.98
1.00
1.37
2.00
2.77
4.00
4.61
L
resp/min
57.6
62.8
50.0
38.6
21.8
R
resp/min
24.8
29.2
29.7
25.9
30.9
L+R
resp/min
82.4
92.0
79.7
64.5
52.7
PRP
rfr/min (s)
0.89
3.73
1.01
3.75
0.98
4.25
1.04
6.14
1.02
8.83
Sessions
22
33
24
20
22
Catania et al., Page 56
Table 2. Least-square fits to left-key data from Experiments 1, 2 and 3.
The fitted equations were exponential, X = Be-KD, hyperbolic, X = B/(1+KD), and logarithmic,
X = AlnD + C, where X is left-key resp/min, D is delay in s, and B, K and C are constants.
Fits
Experiment Pigeons
1
73
(M+N=16)
74
76
78
2
73
(M=8)
74
76
78
3
11
(FI T-s)
54
64
Exponential
r2
B
K
138.8 0.59 0.83
125.0 0.46 0.72
91.4 0.41 0.79
68.6 0.34 0.79
112.2 0.51 0.92
54.3 0.20 0.48
48.9 0.18 0.75
46.8 0.23 0.65
115.3 0.49 0.97
85.6 0.50 0.98
74.8 0.26 0.96
Hyperbolic
r2
B
K
191.3 1.56 0.77
529.2 4.93 0.70
255.8 2.75 0.79
82.8
0.70 0.75
185.4 1.74 0.89
60.5
0.33 0.50
59.9
0.41 0.82
51.9
0.40 0.66
653.9 7.92 0.98
148.0 2.02 0.99
89.6
0.55 0.94
Logarithmic
r2
A x (-1)
C
-44.8
76.2 0.80
-46.2
82.0 0.73
-33.0
62.9 0.79
-21.8
49.5 0.75
-38.4
67.7 0.91
-14.0
46.0 0.50
-13.4
42.2 0.82
-11.7
36.9 0.66
-36.8
69.1 0.95
-26.9
50.6 0.98
-22.0
57.8 0.93
Catania et al., Page 57
Table 3. R-squared estimates for exponential functions in Figure 27.
No fit was available for Pigeon 81, condition Lo. N = number, T = time. Details in text.
Contingency
Pigeon 60
Pigeon 75
Pigeon 77
Pigeon 81
N delay data
O (none)
Hi (drh)
0.488
0.672
0.632
0.110
0.862
0.631
0.436
0.663
Lo (drl)
0.102
0.318
0.333
(none)
T delay data
O (none)
Hi (drh)
0.728
0.476
0.814
0.788
0.934
0.695
0.754
0.803
Lo (drl)
0.717
0.338
0.869
0.608
Catania et al., Page 58
Table 4. Sessions and data for Experiment 5
Reinforcer durations (s), sessions, and maximum reinforcers (rfr) per session, with left (L) and
right (R) response rates shown for the first 20 minutes of sessions and for full sessions for each
pigeon and condition.
Reinforcer
Pigeons duration (s)
3.0
73
4.5
6.0
4.5
3.0
2.0
3.0
4.5
6.0
3.0
Maximum
Sessions rfr/session
14
50
14
35
21
22
21
35
14
50
14
60
28
50
14
35
14
22
14
50
1st 20 min
L
R
resp/min resp/min
42.9
111.5
41.9
105.0
42.5
96.0
36.4
95.9
35.0
101.5
34.8
97.6
40.3
85.9
40.1
87.6
37.2
85.6
31.8
92.5
Session
L
R
resp/min
resp/min
21.7
112.5
29.4
102.6
27.4
93.0
28.8
97.9
21.7
103.3
23.6
95.9
28.6
97.1
27.0
90.6
23.2
86.0
18.8
100.0
74
2.8
4.5
3.0
2.0
3.0
4.5
6.0
3.0
6.0
4.5
3.0
14
14
21
21
14
14
14
14
14
14
14
50
35
50
60
50
35
22
50
22
35
50
30.0
21.0
23.0
22.7
31.3
23.7
22.0
25.5
22.4
27.7
19.4
47.5
50.0
54.9
56.8
45.6
59.3
56.0
55.9
61.6
69.5
61.3
23.1
14.7
17.0
14.4
25.1
13.9
14.7
16.8
13.8
16.6
16.9
49.3
51.6
55.3
56.6
47.7
59.6
56.4
54.3
61.3
66.9
58.3
76
3.0
4.5
6.0
4.5
3.0
2.0
3.0
6.0
3.0
14
35
21
14
14
14
28
14
14
50
35
24
35
50
60
50
22
50
27.4
24.8
22.5
24.3
22.2
23.4
28.1
25.7
27.6
59.8
67.3
67.5
62.9
63.2
66.0
73.4
69.7
64.0
18.4
14.8
14.3
15.8
14.7
14.2
17.1
16.3
19.9
58.3
64.3
68.0
62.3
62.3
69.2
73.0
70.4
63.6
Catania et al., Page 59
Table 5. Effects of varying N with different random-interval (RI) schedules
Required left-key responses = N; delay = time from last left (L) peck to next reinforced right
peck. Details in text.
Pigeon
73
47
N
(right)
2
4
8
4
2
2
4
8
4
2
2
4
8
4
2
RI
(s)
30
2
4
8
4
2
2
4
8
4
2
60
15
60
15
delay
(s)
2.13
2.61
4.94
2.79
2.21
2.31
2.79
4.31
2.38
1.80
1.79
2.14
3.98
2.69
1.91
L
resp/min
22.1
16.8
12.0
19.9
18.8
18.3
20.8
15.3
19.4
23.5
24.7
29.8
19.4
23.0
27.4
0.45
0.63
1.82
0.83
0.50
0.60
0.90
3.04
1.06
0.68
38.1
35.7
20.8
42.8
42.3
30.0
23.8
24.5
22.7
38.7
Pigeon
77
59
N
(right)
2
4
8
4
2
2
4
8
4
2
2
4
8
4
2
RI
(s)
30
2
4
8
4
2
2
4
8
4
2
15
60
15
60
delay
(s)
1.14
1.82
3.06
1.71
0.97
0.97
1.66
3.40
1.64
1.10
0.92
1.55
3.27
1.60
0.93
L
resp/min
22.8
26.5
18.1
21.1
25.1
25.5
19.0
19.5
26.0
26.7
27.4
15.8
11.5
15.1
19.7
Sessions
17
16
25
14
12
17
8
7
9
7
7
7
7
7
7
0.63
1.11
2.30
1.12
0.69
0.69
1.52
2.84
1.41
0.67
51.7
38.0
36.2
42.3
42.7
43.5
40.7
27.3
33.8
51.9
14
12
13
8
10
16
10
14
10
13
Catania et al., Page 60
Table 6. Effects of random interval (RI) on responding maintained by delayed reinforcement
Required left (M) = 8, required right (N) = 4; Max = maximum session duration. Details in text.
Pigeon 60
Pigeon 73
Pigeon 76
RI (s)
30
10
120
60
30
5
90
20
30
15
45
L to rfr
(s)
2.04
2.11
2.70
1.85
1.93
1.64
2.25
1.62
1.75
1.89
1.84
L
resp/min
24.9
15.7
15.3
12.8
17.1
16.2
19.5
27.2
23.7
21.2
26.7
R
Resp/min
47.1
59.6
39.0
54.5
62.7
70.1
52.1
66.9
58.6
70.0
58.4
Max
min
60
30
120
60
60
20
90
30
60
30
60
Sessions
30
19
16
19
14
14
14
14
14
14
21
30
120
60
10
30
90
5
20
30
15
45
3.26
2.82
4.11
5.52
6.62
6.69
*
9.11
9.16
6.99
7.00
17.3
21.1
11.8
4.5
5.8
8.0
*
6.8
9.4
6.9
7.3
51.1
52.6
46.2
40.9
37.4
32.4
*
30.5
30.8
47.6
40.7
60
120
60
30
60
90
20
30
60
30
60
30
22
16
19
14
14
14
14
14
14
21
30
10.8
42.0
60
3.97
30
60
16.2
35.1
60
5.08
20
10
8.9
36.9
30
4.14
15
120
12.6
44.3
120
3.04
19
30
*
*
60
*
14
90
12.5
38.9
90
3.80
14
20
7.0
48.8
30
3.52
14
5
*
*
20
*
14
30
8.1
42.3
60
3.47
14
15
9.9
43.2
30
3.43
14
45
9.1
46.5
60
3.23
21
*Data omitted because of programming/recording problems
Catania et al., Page 61
Appendixes
Appendix 1. Detailed conditions and data from Experiment 1.
Appendix 2. Detailed conditions and data from Experiment 2.
Appendix 3. Detailed conditions and data from Experiment 4.
These Appendixes are provided separately as the pages of an Excel file:
Delay appendixes 1 2 3.xlsx
Figures (Figures 1 through 35) are provided separately in a file named:
Delay Figures.docx
Catania et al., Page 62
Fig. 1. Increases in response rate during 50-s segments of a 500-s fixed interval (FI) in
which alternating segments with light or no light ended with a reinforced response at the end of
the final lighted segment. Periods when the light was off produced low response rates and
therefore interrupted FI responding, but rate when the light was on increased in much the same
way as in a standard FI. (Data averaged across four pigeons; adapted from Dews, 1962, Fig. 2.)
Fig. 2. Schematic of reinforced response sequences. In 1, the last of two responses is
reinforced; in 2, the last of ten is reinforced. Reinforcement probabilities are respectively 1/2 and
1/10, but only two responses are followed by the reinforcer in 1 as opposed to ten responses in 2.
Even allowing for effects of delay, fewer responses are reinforced in 1 than in 2 (cf. Catania,
1971)..
Fig. 3. Schematics of two delayed reinforcement procedures, In the first (top), the delay
resets if a response occurs before it ends. A response, as at x, will always be followed by the
reinforcer after t s, as long as no other response occurs during t. In the second (bottom), the delay
does not reset with responses. After one response initiates the delay, other responses may occur
before the reinforcer is delivered, so that, as at y, the time separating the most recent response
from the reinforcer may be less than t. (Adapted from Catania & Keller, 1981, Figure 4-4.)
Fig. 4. Two hypothetical delay gradients. In both, each response contributes to subsequent
responding in proportion to the height of the gradient at the time it occurs. The contribution in
the top gradient is provided by only a single response. The contribution in the bottom one is the
sum of the contributions provided by each of the several responses.
Fig. 5. Schematic of the topographical tagging procedure, used to separate the respective
contributions to subsequent responding of the peck (b) that precedes the one that produces the
reinforcer and the peck (a) that produces the reinforcer. If the peck that precedes the one that
Catania et al., Page 63
produces the reinforcer must be emitted on Key B while the one producing the reinforcer
continues to be required on Key A, later pecking on Key B can be attributed to the strengthening
of Key-B pecking by a delayed reinforcer, where the delay is given by the time between a and b
(cf. Catania, 1971).
Fig. 6. Illustration of how a delay gradient might operate on responding in a topographical
tagging procedure. A contingency has guaranteed that a sequence of pecks on one key (top) must
be followed by a sequence on a second key (bottom) before a second-key response can produce a
reinforcer. Any response on the first key will be separated (dashed vertical line) from the
reinforcer by these second-key responses, and therefore can only contribute to subsequent
responding in proportion to the height of the delay gradient at the time at which it occurs.
Fig. 7. Data from four pigeons in a topographical tagging procedure; the smooth curves
show exponential fits. Variable-interval (VI) schedules arranged reinforcers for sequences
consisting of one left-key peck followed by N right-key pecks, with N ranging from 1 to 11.
Left-key response rates are plotted against the estimated mean time between the left-key peck
and the reinforcer produced by the last peck of the right-key sequence (mean B-key IRT x N;
mean changeover times from B to A were unavailable). The plotted data are derived from
Catania (1971), Table 2.
Fig. 8. Two hypothetical gradients. Some procedures use measures that effectively
determine the height of the gradient at a given delay of reinforcement (vertical line, top),
whereas others use measures that allow estimates of the area under the gradient up to that delay
(filled area, bottom). Details in text.
Catania et al., Page 64
Fig. 9. Schematic example of a delay gradient generated by a reinforcer (rfrx), at the top,
and the truncation of that gradient by the introduction of an earlier reinforcer (rfry), at the
bottom. Details in text.
Fig. 10. Standard pigeon key equipped with a Hall Effect switch instead of a make-break
electrical contact. Details in text.
Fig. 11. Schematic of responding on left (L) and right (R) keys during the time between two
reinforcers (RFR) in a segment of a topographical-tagging session with random-interval (RI)
scheduling of reinforcers (PRP = postreinforcer pause). Details in text.
Fig. 12. Schematic illustration of consequence of the traditional method of arranging RI
schedules and the method used in the present studies. In the former, timing stops with
uncollected setups; in the latter, timing continues and setups can accumulate. The difference
between obtained and scheduled reinforcers is smaller in the latter than in the former. Details in
text.
Fig. 13. Responses per minute on the left key as a function of the time between the last left
peck and the later reinforced right peck, for four pigeons. The lines are exponential fits to the
data.
Fig. 14. Overall response rate (sum of left-key and right-key rates) as a function of the
number of pecks required on the left key (M, where M plus N equals sixteen and N is the number
of pecks required on the right key). The lines are linear fits to the data.
Fig. 15. Responses per run on the left and right keys as a function of the required pecks on
the right key (N). The lines are linear fits to the data.
Catania et al., Page 65
Fig. 16. Time from the last left peck to the next reinforced right peck (delay of
reinforcement) as a function of the required pecks on the right key (N). The lines are exponential
fits to the data.
Fig. 17. Changeovers (COs) and postreinforcement pauses (PRPs) as a function of the time
from the last left peck to the next reinforced right peck (delay of reinforcement). The lines are
linear fits to the CO data only.
Fig 18. Rate of responding on each key as a function of the time from the last left peck to
the next reinforced right peck (delay of reinforcement). The lines for the left-key rates are
exponential fits. Those for the right-key data, not well fit by exponential functions, are
logarithmic fits.
Fig. 19. Responses per minute on the left key as a function of the number of pecks required
on the left key (M), with the required pecks on the right key (N) held constant at 8, for two
pigeons. The lines are linear fits to the data.
Fig. 20. Time from the last left peck to the next reinforced right peck (delay of
reinforcement) as a function of the required pecks on the right key (N), in Experiment 2. The
lines are exponential fits to the data. Cf. Figure 16.
Fig. 21. Responses per minute on the left key as a function of the time between the last left
peck and the later reinforced right peck, for four pigeons in Experiment 2. The lines are
exponential fits to the data. Cf. Figure 13.
Fig. 22. Total rates of responding (left key plus right key) and right-key rates as a function
of the time from the last left peck to the next reinforced right peck in Experiment 2. The fits are
linear for the former and logarithmic for the latter. Details in text, and cf. Figures 14 and 18.
Catania et al., Page 66
Fig. 23. Responses per minute on the left key as a function of the time between the last left
peck and the later reinforced right peck, for three pigeons in Experiment 3. The lines are
exponential fits to the data. Cf. Figures 13 and 21.
Fig. 24. Total rates of responding (left key plus right key) and right-key rates as a function
of the time from the last left peck to the next reinforced right peck in Experiment 3. Details in
text, and cf. Figure 22.
Fig. 25. Schematic of the possible effects of rate shaping based on two different
assumptions: that the effects of delay depend solely on the time separating behavior from a later
reinforcer (time delay, left column) or that they depend solely on the number of responses
separating behavior from a later reinforcer (number delay, right column). The schematic shows
different contingencies arranged for right-key runs and illustrates the contrasting effects that
delays would produce if based on one or the other of these assumptions. Details in text.
Fig. 26. Three perspectives on an XYZ plot of data for Pigeon 60. With the overall relation
shown in the center, the left graph emphasizes the relation between number delay and left-key
responses rates and the right graph emphasizes the relation between time delay and left-key
responses rates. Details in text.
Fig. 27. Left-key response rates plotted against number delays (left column) and time delays
(right column) over various contingencies that shaped lengths of right-key runs and local rightkey response rates for four pigeons (O = none, Hi = drh, Lo = drl). Fits are logarithmic. Details
in text.
Fig. 28. Three perspectives on XYZ coordinates showing how a single function relating
response rate to number delay and time delay (filled circles within each graph) can be compatible
Catania et al., Page 67
with a function totally determined by variable X (row A), or with one totally determined by
variable Y (row B), or with one jointly determined by both X and Y (row C).
Fig. 29. Data in XYZ coordinates for Pigeon 81, showing data the variation of which with
number delays and time delays resembles the hypothetical data with filled circles in Figure 28.
Symbols and conditions are as in Figure 26.
Fig. 30. Points in the number-delay time-delay coordinate space (the base of the XYZ
graphs of Figures 26 and 29) explored with each pigeon during Experiment 4.
Fig. 31. Hypothetical delay gradients showing how the truncation of gradients as one
reinforcer moves closer in time to another (as in B through E) can lead to a gradient engendered
by two reinforcers (E) that is essentially equivalent to one engendered by a single reinforcer (A).
Details in text.
Fig. 32. Left-key response rates maintained as a function of reinforcer durations for four
pigeons. Data are shown both for the first twenty min of sessions and for full sessions to
accommodate possible effects of within-session changes in response rates. The lines provide
linear fits.
Fig. 33. Response rates as a function of time to the reinforcer with the value of the RI
schedule as a parameter, for four pigeons. The data are shown with exponential fits.
Fig. 34. Left-key and right-key response rates as a function of the value of the RI schedule
(left column), and delays from the last left peck to the next reinforced right peck (time delay) as
a function of the value of the RI schedule (right column), for three pigeons. Details in text.
Fig. 35. Observing-response rates as a function of the time to a reinforcer at the end of a
fixed-interval, with exponential fits for three pigeons. (Adapted from Catania, 2012, Figure 2).
Download