Uploaded by hannahegill07

blpquizlet

advertisement
BLP Exam 2
Study online at quizlet.com/_4lvo79
1.
Abuse
Liability of
Drugs
"Abuse liability" of drugs is related to
"reinforcing efficacy"
6.
Belongingness
Keller and Maria Breland
-both were Skinner students
-Worked in Hollywood training animals to
perform certain behaviors for films
-Some behaviors could not be trained
because they were too alike the animal's
natural tendencies. They belonged.
7.
B.F Skinner
A behaviorist who studied operant
conditioning and schedules of reinforcement
Wrote 'The Behavior of Animals'
8.
Break and Run
Pattern
You get your reinforcement for having done
it once then you realize you have to do it
again so you take a break, then you do it
again.
Break is caused by anticipation of more
work ahead
Break is longer if you anticipate larger
amount of work
9.
Break Point
When the animal will not complete the
response any longer after ratio strain.
10.
Continuous
Reinforcement
Reinforcement is delivered after every
response.
Also known as FR1
Joe Brady
-A Skinner student
-Worked with rats, primates, and human
volunteers
-Ratio strain and "break point"
-Fixed Ratio Schedules
-Progressive Ratio Schedules
2.
According to
Thorndike,
what causes
learning to
occur?
Reinforcement or punishment after a response
3.
Annoyers
This that are agitating ---these lead to weaker
responses when paired with a stimulus.
In emitted behavior!
4.
Appetitive
Stimuli
The "satisfying" stimulus
Net effect on behavior does not depend on
the type of stimulus alone (appetitive versus
aversive).
Example: Child gets rewarded every time he
or she does a desired behavior (maybe a
gold star for every perfect spelling score)
The type of contingency between the
response and stimulus outcome interacts with
the kind of outcome.
11.
That is, both positive reinforcement (involving
an appetitive stimulus) and negative
reinforcement (involving an aversive stimulus)
strengthen behavior, that is, they are both
reinforcing.
5.
Aversive
Simuli
The "annoying" stimulus
Net effect on behavior does not depend on
the type of stimulus alone (appetitive versus
aversive).
The type of contingency between the
response and stimulus outcome interacts with
the kind of outcome.
That is, both positive reinforcement (involving
an appetitive stimulus) and negative
reinforcement (involving an aversive stimulus)
strengthen behavior, that is, they are both
reinforcing.
Cumulative
Record
A device created by Skinner that record the
responses and rewards through time.
It produces a graphical record of the
accumulation of responses over time.
A steep slope means there is a fast rate of
response, likewise, a shallow slope means
there is not a fast rate of response.
A flat line means there is no response and a
slash in the paper is an indication of
reinforcement.
Distance on X axis shows how much time
has passed.
Height on Y axis shows how many
responses there are.
12.
Describe the
development of the
generalization gradient
as training progresses.
It gets steeper. Discrimination
training makes the generalization
gradient even more steep.
13.
Differential
Reinforcement of Low
Rates (DRL) Schedule
Like fixed interval schedule,
except early responses reset the
time interval (reinforcement is
delayed)
Produces very low rates of
responding
19.
Escape
and
Avoidance
Procedure
aka negative reinforcement
20.
FI Scallop
The gradually increasing rate of responding that
occurs between successive reinforcements on a
fixed-interval schedule.
21.
Fixed
Interval
A reinforcement schedule in which the first
correct response after a set amount of time has
passed is reinforced
The time period required is fixed (always the
same).
Is sensitive to deficits in internal
timing, such as produced by
alcoholic Korsakoff's disease
14.
Discrete Trial Procedures
15.
Drug Self Administration
16.
Elicited Behavior
Behavior that is drawn out by a
stimulus, not everyday behavior.
17.
E.L Thorndike
Studied instrumental conditioning
in cats in "puzzle boxes"
Proposed two ways of learning:
1. Trial and Error
2. Insight
Emitted Behavior
22.
Fixed Ratio
Animal controls when to get the
injection
• Or can stick a tube into a
specific part of the brain and
inject cocaine and you will get
strong reinforcement
Want to find out HOW
reinforcing is the drug
• Reinforcement is directly
correlated with abuse
o Animals will ignore food
because the drug is so
reinforcing, they can starve to
death
18.
Example: putting clothes in the washer then you
begun studying. After a while you know it is
around the time to check to see if the clothes
are ready. If not, then you come back again
later when you think it is closer to the time
again.
Each trial is "discrete", which
means each trial has a beginning
and an end.
There are two types:
Latent - response time
Correct vs. Incorrect Response
Spontaneous behavior!
Cannot always predict the
stimulus making the behavior.
Comes from within!
ex.) Bella and Sophie running
about in the backyard
An instrument conditioning procedure in which
the instrumental response terminates the
aversive stimulus
A reinforcer is administered after a specified
number of correct responses.
THIS IS BEST FOR LEARNING A NEW
BEHAVIOR
Ex.) FR210, FR900
only delivered after 210 responses or 900
responses
23.
Flat
Gradient
When the animals responds the same to all the
responses
24.
Free
Operant
Procedures
Free Operant allows the participant to respond
whenever they want and hundreds or thousands
of times.
Conditions measured:
-Most common type: the rate (the number of
responses per unit of time)
-The changes in interresponse time (the time in
between responses)
-The changes in choice probability (ex. between
two or more response levers)
25.
Guttman & Kalish's
(1956) Experiment
• Trained pigeons to peck when a 580
nm light illuminated a key (yelloworange)
• Then, trained them on partial
reinforcement
• Then "probed" with nonreinforced trials
with different light wavelengths
• Most responses occurred for probes
that had the same color wavelength as
training
• The more a probe differed from the
training stimulus (the color yelloworange), the fewer the responses
26.
27.
28.
Honig et al. (1963)
How does the
phenomenon of
learned helplessness
demonstrate that
people and animals
can learn "no
contingency"? How
can learned
helplessness be
overcome?
In the Seligman and Maier
experiment, they had a group of
dogs (Group Y) that were unable to
escape the shocks they were
administered no matter that their
response was. Eventually these
dogs gave up and accepted the
shocks. They had no contingency.
30.
How does the Spence
model explain peak
shift?
Inhibition caused by the S- cancels
out some of the excitation caused
by the S+, so the peak appears to
shift.
31.
How do you know
which direction in the
stimulus dimension the
peak will shift, up or
down the stimulus
dimension?
The peak for the S+ generalization
gradient shifts in the direction "away
from the S-."
32.
Insight Learning
An "aha!" moment. An epiphany.
29.
Used line tilt stimulus
-Trained pigeons to peck a vertical
black line (90 degrees) and presented
them with other lies at varying degrees
(30, 60, 120, 150,180)
-the pigeons pecked more at the lines
that were similar to 90 (60/120) than the
ones that weren't that close to vertical.
How are ratio
strain and the
break point used
to assess the
abuse liability of
drugs?
...
How did Honig et
al.'s (1963)
experiment
demonstrate that
discrimination
training can cause
stimuli to acquire
excitatory and
inhibitory
properties?
The peak shift.
When discrimination was learned, the
peak shifted away from the original
learned stimulus. The original stimulus is
excitatory and the discriminant stimulus
is the inhibited stimulus.
They dogs overcame this when
they were shown by people that
they could now escape the shocks.
One of the two ways you could
learn according to Edward
Thorndike
33.
Instinctive Drift
A behavior MUST conform to the
animal's behavior
ex.) Keller and Maria Breland were
trying to train raccoons to take a
coin and place it inside of a box.
After the raccoons would do it
correctly, they were given a food
reward. -However, the raccoon
became classically conditioned,
and associated the coin and food.
-Raccoons clean their food (elicited
behavior)
-Eventually the raccoons began to
take the coin and clean it in nearby
water like they would food
because they anticipated the food.
-This could never be broken
because you cannot override a
reflex.
34.
Jenkins &
Harrison
(1962)
Experiment
Some pigeons were trained to peck for white
noise, but got no discrimination training
-They responded equally to all probes
Other pigeons were trained to peck for a 1000
Hz tone, but got no discrimination training (S+,
no S- group).
-A typical generalization gradient was
observed with peak responding at the training
stimulus.
Other pigeons were trained to peck for a 1000
Hz tone, but not to peck for 950 Hz (S+,
Sgroup).
-A steeper generalization gradient was
observed
40.
Learned
Helplessness
Seligman Dog Experiment
41.
Learning by
Consequences
42.
Lieberman,
McIntosh, &
Thomas' (1979)
RESULTS:
Discrimination training improves
discrimination.
Peak Shift was observed
35.
Joseph E.
Brady
Studied the abuse potential of drugs with
schedules of reinforcement
36.
Keller and
Maria
Breland
Skinner students who worked in the movie
industry and demonstrated "instinctive drift" in a
variety of animals they trained, including
raccoons
Founded Animal Behavior Enterprises in 1943
which they used to train animals for Hollywood
films
Korsakoff's
Disease
A chronic memory disorder causes by a severe
lack of vitamin B due to excessive drinking
38.
K.W
Spence
Proposed the Spence model of discrimination
learning
39.
Law of
Effect
Proposed by Thorndike
If a response in the presence of a stimulus (S) is
followed by a satisfying event, the association
between the stimulus and the response (R) will
be strengthened.
• The reinforcement causes the learning!
• Example: If, in the puzzle box (S), the cat
pushes a lever (R) and then gets food (a
satisfying event), the cat will learn to make the
response (R) whenever it is in the puzzle box
(S), that is, the S-R association is strengthened.
o This is a positive relationship
AKA positive reinforcement
• Random emitted behavior is strengthened
or weakened by its consequences
• Consequences: satisfying versus annoying
events
• Training rats to go into a certain color box
in order to earn a reward.
• The interval is 1 minute between the color
box and receiving the reward
• The minute interval is too long for the rats
to learn to go into a certain color box
• How can we make the rats learn?
o Marking the correct behavior
• If the rats go into the correct box, they are
picked out of the box (the mark) and put
into the delay box
o This makes the minute interval irrelevant
to learning now
• This means you can have a long delay
interval for reinforcement if there is
something to mark the correct behavior
o Marking forms the connection
They worked with Skinner on the WWII Pigeon
Guided Bombs
37.
When exposure to inescapable and
unavoidable aversive stimulation reduces
motivation to respond and disrupts
subsequent instrumental conditioning
because participants learn that their
behavior does not effect the outcome.
43.
Mark Bouton
Discovered renewal and reinstatement
44.
Marking
Preliminary stage to the instrumental
conditioning process.
Pairing the stimulus to reinforcer so the
participant learns to go to the reinforcer
when it is presented.
ex.) pairing the sound of food delivery to
food so an animals knows to go to the food
bowl when they are to receive the reinforcer.
45.
Martin E.P
Seligman
Studied "learned helplessness" in dogs
46.
The
Misbehavior of
Animals
Keller and Maria Breland wrote this
-Described the things you could not do
with animals because they would not work
-A slap to Skinner because they were his
students and he came out with a book
beforehand called "The Behavior of
Animals"
47.
Negative Law
of Effect
Proposed by Thorndike
• If a response in the presence of a stimulus
(S) is followed by an annoying event, the
association between the stimulus and the
response (R) will be weakened.
• Example: If, in the puzzle box (S), the cat
pushes a lever (R) and then gets a shock
(an annoying event), the cat will learn to
withhold the response (R) whenever it is in
the puzzle box (S), that is, the S-R
association is weakened
o This is a negative relationship
AKA positive punishment!
48.
Negative
Punishment
51.
Positive
Reinforcement
An instrumental conditioning procedure in
which there is a positive contingency
between the instrumental response and an
appetitive stimulus or reinforcer.
-If the participants performs the response, it
receives the reinforcer if it does not, it will
not get the reinforcer (the appetitive
stimulus)
52.
PostReinforcement
Pause
The pause after the reinforcement is
delivered and before the beginning of the
next "run" phase.
This pause is longer if the anticipated work
load is large
53.
Progressive
Ratio
Schedules
A schedule made up of a series of
increasing fixed ratios.
An example PR schedule: FR1, FR2, FR4,
FR6, FR9, FR12, FR15, FR20, FR25, etc.
PR schedules can test the limits of the
reinforcing properties of the reinforcer (i.e.,
how hard an animal will work for a given
reinforcer).
Also known as Omission Training, Time Out,
and Differential Reinforcement of Other
Behavior (DRO)
An instrumental conditioning procedure in
which a positive reinforcer is periodically
delivered when the target response is not
performed.
54.
Punishment
An instrumental conditioning procedure in
which there is a positive contingency
between the instrumental response and an
aversive response.
-Prevents the subject from performing the
target response.
49.
Negative
Reinforcement
ex.) a shock
50.
Partial
Reinforcement
If the participant performs the instrumental
response, it receives the aversive stimulus, if
the participant does not perform the
instrumental response, they will not get the
aversive stimulus.
An instrumental conditioning procedure in
which there is a negative contingency
between the response and the aversive
stimulus (or reinforcer).
-If the response is performed, the aversive
stimulus is cancelled, however if the
response is not performed it will get the
aversive stimulus.
A schedule of reinforcement in which only
some of the occurrences of the instrumental
response are reinforced. The response is
only reinforced occasionally, or
intermittently.
Also known as positive punishment
55.
Puzzle Box
Experiment
Edward Thorndike created puzzle boxes
with slats and placed cats inside of them
-he then placed a bowl of cat food on top
of the box
-the cats would grow hungry and try and fit
their paws in between the slats to try and
get the food but they could not reach
-eventually the cats would habituate and
they would forget
-cats would begin to explore
-would accidentally pull the pulley/lever,
etc. and become free
-cats would dishabituate and go get the
food
-Thorndike was looking to investigate
higher level cognition
-Cats would not always realize they had to
pull the lever
56.
Ratio Strain
When the amount of work is too much for
the small amount of reinforcement the
animal receives.
Animal may pause before completing the
whole response.
57.
Reinstatement
Reinstatement is different from spontaneous
recovery because reinstatement is when
you are triggered by a US versus
spontaneous recovery which is when you
go a period of time without receiving the
stimulus then you get it and you exhibit the
response again.
Reinstatement: recovery of excitatory
responding to an extinguished stimulus
produced by exposure to the US
58.
Renewal
...
59.
Response
Rate
Indicated by the slope
The slope of the cumulative record
indicates the response rate measured as
responses per unit time (e.g., responses per
second)
Response rate is a sensitive measure of
conditioning
60.
Satisfiers
65.
66.
Stimulus
Discrimination
Stimulus discrimination is differential
responding between two or more stimuli.
Discrimination learning speeds up
learning!
67.
Stimulus
generalization
and the
stimulus
generalization
gradient
The more you differ from the trained
stimulus, the fewer responses you get.
Trial-and-Error
Learning
As it implies, it is trial and error. The cats in
the puzzle box experiment do not always
get the behavior correctly.
68.
These are pleasurable. These lead to
stronger responses when paired with a
stimulus.
Schedules of
Reinforcement
A rule that determines when reinforcement
will be delivered
62.
Shaping
(Skinner)
A complex behavior can be "shaped" with a
series of changing contingencies called a
program
Each stage evokes a response and also
prepares the organism to respond at a later
stage.
- Response "chaining
64.
Spence Model
of
Discrimination
Learning
Can only memorize stimuli
-cannot learn abstract rules such as
relational learning
-animals that fall under this are pigeons
Steep
Gradient
When the animal learns to respond to the
stimulus they are trained to and the ones
VERY similar to it.
The more a probe differed from the training
stimulus, the fewer the responses
In reference to the Guttman and Kalish
experiment with the pigeons and color
gradient.
One of the two ways you could learn
according to Thorndike
69.
61.
Responding may come under the control of
the stimuli that are present when the
response is reinforced. The association
between the stimulus and the response is
strengthened by the reinforcer.
Stimulus control is demonstrated by
stimulus discrimination rather than stimulus
generalization.
In emitted behavior!
63.
Stimulus
Control
Variable
Interval
The first correct response after a set amount
of time has passed is reinforced, but the
time period changes constantly.
After the reinforcement, a new time period
(shorter or longer) is set with the average
equaling a specific number over a set of
trials.
Example: checking social media versus
checking your email. If you get more social
media notifications, then you will check it
more often than the email.
70.
Variable Ratio
A reinforcement is given after a set number
of correct responses.
After reinforcement is administered, the set
number of correct responses for future
reinforcement changes again.
This is best for MAINTAINING behavior!
ex.) gambling
71.
What causes the post-reinforcement
pause, ratio strain, and break point
observed in fixed ratio schedules?
Procrastination
72.
What disease did we discuss in class that
has been studied using one of these
schedules?
Alcoholic
Korsakoff's
disease
73.
What effect does positive punishment
have on behavior? What effect does
"negative punishment" have on behavior?
What's another everyday name for
"negative punishment"?
[Hint: They are both forms of punishment,
so they both cause what effect on
behavior...?]
They weaken a
behavior
What effect does positive reinforcement
have on behavior? What effect does
negative reinforcement have on behavior?
[Hint: They are both forms of
reinforcement, so they both cause what
effect on behavior...?]
They both
strengthen a
behavior
74.
Other names
for Negative
Punishment is
1. Omission
Training
2. Time Out
3. DRO
75.
What is peak
shift?
The peak of the generalization gradient
appears to shift.
The observed response is the sum of
these gradients. (yellow)
76.
What is "relational
learning"?
It's learning a general rule or
relationship between multiple stimuli
instead of memorizing what to do for
each specific stimulus.
-primates and rats do this!
77.
What is shaping?
The formation of a behavior through
reinforcement and "chaining"
Chaining is gradually connecting
multiple behaviors together to create a
"chain"
78.
What is the
difference
between
appetitive and
aversive stimuli?
Appetitive is satisfying and aversive is
annoying
79.
What is the
difference
between
discrete-trial and
free-operant
procedures?
Discrete trials have a fixed number of
time to complete the trial
What is the
difference
between Elicited
and Emitted
behavior?
Elicited is drawn out by a known
stimulus (a US or CS)
What is the
relationship of the
steepness of the
stimulus
generalization
gradient and
stimulus control?
The more training occurs, the steeper
the generalization gradient.
aka the more the animal learns a
stimulus, the less they will react to a
stimulus that isn't the trained one.
80.
81.
In Free operant, there is no time
constraint and the experimenter does
not intervene
Emitted comes from within, you cannot
tell what the stimulus is that is causing
the behavior
82.
What schedules require especially good response
timing?
FI and DRL schedules
83.
When you consider the experiment of Lieberman et
al. (1979), why do you think a longer delay of reward
causes poorer learning?
With longer delays, it's hard to identify the contingency between behavior
and the outcome.
84.
WWII: "Project Pigeon" and Pigeon-Guided Bombs
Skinner trained pigeons to peck at a video image to hold a missile onto a
target.
• Tried to use pigeons as a guidance system for the bombs.
• Put pigeons in a machine and the screens were connected to a camera
that looked forward and if a pigeon saw a target, they would peck at it
which would direct the fin of the drone.
• It worked.
Download