xan-XAN-2014-1122-OnLine_for_PsycARTICLES

advertisement
Appendix for the APA Website
The Rescorla-Wagner model of classical conditioning (Rescorla and Wagner 1972; for a review, see
Miller, Barnet and Grahame 1995) assumes that the gain in associative strength due to the pairing of
a CS with a US is proportional to the difference between the maximum strength of association that
the US can support, and the current associative strength of all the CSs present on this trial. As a
consequence, the rate of change in conditioned responding to a CS decreases as the number of
training trials increases, resulting in a negatively accelerated curve of learning across training.
More formally, two equations describe how associative strength of a given CS (CSx) change on each
trial:
∆𝑉𝑋𝑛+1 = 𝑎𝑋 𝛽(𝜆 − 𝑉𝑡𝑜𝑡 )
(1)
and
𝑉𝑋𝑛+1 = 𝑉𝑋𝑛 + ∆𝑉𝑋𝑛+1
(2)
where ∆𝑉𝑋𝑛+1 is the change in the associative strength (V) of CSx on Trial n + 1; αx is the associability
of CSx (range: 0 to 1), β is the associability of the US (range: 0 to 1); λ is the maximum associative
strength that the US can support; and Vtot is the sum of associative strengths of all CSs (including X )
present on Trial n. In Equation 2 , 𝑉𝑋𝑛+1 is the associative strength of CSx after Trial n + 1; 𝑉𝑋𝑛 is the
associative strength of CSx immediately before Trial n + 1 (i.e. , after trial n); and ∆𝑉𝑋𝑛+1 is the change
in the associate strength of CSx as a result of Trial n + 1, as assessed by Equation 1.
The training sequences were constructed by drawing randomly among a set of runs from which
number and length were previously determined. For each type of trials (i.e., CS-US and CS-alone),
there were one run of five trials, two runs of four trials, four runs of three trials, eight runs of two
trials and 16 runs of a single trial. The resulting sequences comprised 114 trials [i.e.,
2(1*5+2*4+4*3+8*2+16*1)] and conformed exactly to a binomial distribution of two equally
probable events.
In the Perruchet's paradigm, there is a single CS and a single US. In the simulation below, αβ was
given the function of a learning rate parameter. The paradigm comprises both CS-US pairs
(acquisition) and CS alone trials (extinction). λ was set to 1 for acquisition trials and zero for
extinction trials, and αβ took different values for the different type of trials. The Rescorla-Wagner
model assumes that αβ for extinction is a number smaller than αβ for acquisition but larger than
zero. αβ was set to 0.04 for CS-US pairs and 0.03 for CS-alone trials in Simulation (A), and these
values were increased exponentially from Simulation (A) to simulation (D). These values were
empirically selected to cover cases in which the learning curve does not reach asymptote before the
end of the 114 trials (A), to cases in which asymptote is reached within the very first trials (D).
Simulation (E) was aimed to test whether the assumption that αβ for extinction is a number smaller
than αβ for acquisition is crucial, by inverting the relation. Finally, Simulation (F) was aimed at
exploring whether the ratio between the learning rate parameters for acquisition and extinction was
important, by using much more distant values (see Table 1).
Values of αβ
Simulations E1-E2 trials E1-alone trials
0,04
0,03
(A)
0,08
0,06
(B)
0,16
0,12
(C)
0,32
0,24
(D)
0,06
0,08
(E)
0,16
0,03
(F)
TABLE 1.
The results shown in Figure 1 have been averaged over 10.000 randomly generated sequences for
each simulation. Unsurprisingly, associative strength (upper panels) increased following a negatively
accelerated function across the 114 trials for all simulations. The value of the asymptote and the
number of trials needed to reach this asymptote differed in a predictive way as a function of learning
parameters. Except for Simulation (F) in which there was a strong asymmetry between the leaning
rates for acquisition and extinction, the asymptotic value is far from 1 (λ), which is typical of the
prediction of the model with a partial reinforcement paradigm.
The variation of associative strength as a function of the length and type of the preceding run of trials
was more important for our concern. An overall look on the lower panels of Figure 1 suggests the
prevalence of an upward linear trend. However, there were two minor exceptions. First, there was a
slight, but consistent departure from linearity in the middle part of the curves. Second, the curve for
Simulation (D) tended to be sigmoïdally shaped. Recall that the learning rates were very strong,
leading to reach asymptote within the very first trials (around eight trials, i.e. four E1-E2 trials on the
average). Neither of these exceptions violates monotonicity. These exceptions could turn out to be
relevant for further fine-grained analyses of the data, but for any practical purpose, it appears that a
linear function provides a satisfactory approximation for making predictions.
Figure 2
Download