PSY105 Neural Networks 2/5

advertisement
PSY105 Neural Networks 4/5
4. “Traces in time”
Assignment note: you don't need to read the full book to
answer the first half of the question. You should be able to
answer it based on chapter 1 and the lecture notes.
Lecture 1 recap
• We can describe patterns at one level of
description that emerge due to rules followed
at a lower level of description.
• Neural network modellers hope that we can
understand behaviour by creating models of
networks of artificial neurons.
Lecture 2 recap
• Simple model neurons
– Transmit a signal of (or between) 0 and 1
– Receive information from other neurons
– Weight this information
• Can be used to perform any computation
Lecture 3 recap
• Classical conditioning is a simple form of
learning which can be understood as an
increase in the weight (‘associative strength’)
between two stimuli (one of which is
associated with an ‘unconditioned response’)
Nota Bene
• Our discussion of classical conditioning has
involved
– A behaviour: learning to associate a response with
a stimulus
– A mechanism: neurons which transmit signals
– These are related by…a rule or algorithm
Learning Rules
“When an axon of cell A is near enough to excite
a cell B and repeatedly or persistently takes
part in firing it, some growth process or
metabolic change takes place in one or both
cells such that A's efficiency, as one of the
cells firing B, is increased.”
Hebb, D.O. (1949), The organization of behavior,
New York: Wiley
Operationalising the Hebb Rule
• Turn ….“When an axon of cell A is near enough
to excite a cell B and repeatedly or persistently
takes part in firing it, some growth process or
metabolic change takes place in one or both
cells such that A's efficiency, as one of the
cells firing B, is increased.”
• ….Into a simple equation which is a rule for
changing weights according to inputs and
outputs
A Hebb Rule
• Δ weight = activity A x activity B x learning rate
constant
• In words: increase the weight in proportion to the
activity of neuron A multiplied by the activity of
neuron B
Stimulus Off
Stimulus On
Time
Implications of this rule
CS1
Stimulus 1
?
Stimulus 2
UCS
Implications of this rule
CS1
Stimulus 1
?
Stimulus 2
UCS
Implications of this rule
CS1
Stimulus 1
?
Stimulus 2
weight
UCS
Implications of this rule
Activity A
CS1
Stimulus 1
?
weight
Activity B
Stimulus 2
UCS
Implications of this rule
Activity A
CS1
Stimulus 1
?
weight
Activity B
Stimulus 2
=
x
UCS
x 0.1
The most successful model of Classical
Conditioning is the Rescorla Wagnar
model
Accounts for the effects of combinations of stimuli in
learning S-S links
Based on the discrepancy between what is expected to
happen and what happens
But…
Deals with discrete trials…ie has no model of time
Rescorla RA, Wagner AR. A theory of Pavlovian conditioning: Variations in the effectiveness of reinforcement and
nonreinforcement. In: Classical Conditioning II: Current Research and Theory (Eds Black AH, Prokasy WF) New York:
Appleton Century Crofts, pp. 64-99, 1972
Robert Rescorla (2008) Rescorla-Wagner model. Scholarpedia, 3(3):2237.
The problem of continuous time
Stimulus 1
Stimulus 2
The problem of continuous time
Stimulus 1
Stimulus 2
The problem of continuous time
Stimulus 1
Stimulus 2
The problem of continuous time
Activity A
Stimulus 1
Activity B = 0
Stimulus 2
The problem of continuous time
Activity A
Stimulus 1
Activity B = 0
Stimulus 2
The problem of continuous time
Activity A = 0
Stimulus 1
Activity B
Stimulus 2
• We need to add something to our model to
deal with a learning mechanism that is always
“on”
Traces
Stimulus 1
Traces
Stimulus 1
Traces
Stimulus 1
Stimulus 2
Traces
Activity A
Stimulus 1
Activity B
Stimulus 2
Traces
Stimulus 1
Stimulus 2
Consequences of this implementation
•
•
•
•
•
•
Size of CS stimulus
Duration of CS stimulus
Size of UCS stimulus
Duration of UCS stimulus
Separation in time of CS and UCS
The order in which the CS and UCS occur
– (cf. Rescola-Wagner discrete time model)
Download