Chapter 13 The Behavioral/Social Learning Approach: Theory and Application Behaviorism John B. Watson (1878-1958) Watson was a fighter and a builder. In college, he was unsociable and uninterested in his studies. He switched from philosophy to psychology at the University of Chicago, but preferred to study animals rather than humans. After joining the faculty at Johns Hopkins University in 1908, he developed and promoted his ideas about behaviorism. Behaviorism swept the field following the publication of his paper “Psychology as the Behaviorist Views It” in 1913. Watson’s flourishing academic career ended in disgrace in 1920, in the wake of his affair with his research associate, Rosalie Rayner. Watson’s view of behaviorism Watson argued that if psychology were to be a science, psychologists must stop examining mental states and study overt, observable behavior instead. Emotions, thoughts, experiences, values, reasoning, insight, and the unconscious would therefore be off-limits to behaviorists unless they could be defined in terms of observable behaviors. For example, Watson regarded thinking as a variant of verbal behavior that he called “subvocal speech.” According to Watson, personality could be described as “the end product of our habit systems.” Watson argued that he could take “a dozen healthy infants, wellformed” and condition them “to become any type of specialist” that he chose. B.F. Skinner (1904-1990) Skinner passed on joining his father’s law firm and studied English at Hamilton College with the intent of becoming a writer. After producing nothing of consequence in the two years following his graduation, he went to Harvard to study psychology. He became the new standard bearer for a view of behaviorism that he called radical behaviorism. In the 1940s, he published Walden Two, a novel about a utopian community based on conditioning principles. To the end, he remaining an adamant believer in the power of the environment, and conceded little to those who emphasized genetic determinants of behavior. Skinner’s “radical behaviorism” Skinner did not deny the existence of thoughts and inner experiences, but he argued that we often misattribute our actions to mental states when they should be attributed to our conditioning instead. His position maintained that we often don’t know the reason for many of our behaviors, although we may think we do. He therefore argued that our perception that we are free to act as we choose is, to a large extent, an illusion. Instead, we act in response to environmental contingencies. Among Skinner’s many contributions to the study of conditioning are the so-called “Skinner box” and the discovery of partial reinforcement schedules. Basic Principles of Conditioning Ivan Pavlov (1849-1936) Ivan Pavlov was born in Ryazan, Russia. He began his higher education as a seminary student, but dropped out and enrolled at the University of Petersburg to study the natural sciences. He received his doctorate in 1879. In the 1890s, Pavlov was investigating the digestive process in dogs by externalizing a salivary gland so he could collect, measure, and analyze the saliva produced in response to food under different conditions. He noticed that the dogs tended to salivate before food was actually delivered to their mouths. He realized that this was more interesting than the chemistry of saliva, and changed the focus of his research, carrying out a long series of experiments in which he manipulated the stimuli occurring before the presentation of food. He thereby established the basic laws for the establishment and extinction of what he called "conditional reflexes" — i.e., reflex responses, like salivation, that only occurred conditional upon specific previous experiences of the animal. Pavlov’s experimental setup Classical (Pavlovian) conditioning (aka “signal learning”) Classical conditioning begins with an existing stimulus- response (S-R) association. Understanding that there was an existing S-R association between the food (S) and the dog’s salivation, Pavlov quickly perceived that there might also be a learned or “conditioned” association between cues associated with feeding (S) and the dog’s salivation (R). Using the sound of either a bell or a tuning fork as his conditioned stimuli, Pavlov found that he could indeed “condition” the response of salivation to the sound of a bell or a tuning fork. Pavlov’s experimental setup Pavlov’s famous demonstration of classically conditioned salivation in a dog Classical conditioning paradigm food Unconditioned stimulus (UCS) Conditioned stimulus (CS) tuning fork saliva Unconditioned response (UCR) Conditioned response (CR) saliva Classical (Pavlovian) conditioning (aka “signal learning”) Once the new S-R association is established, it can be used to condition yet another S-R association in a process called second-order conditioning. For example, once the dog is reliably salivating to the sound of the tuning fork, the tuning fork can be paired with a green light and soon the dog will salivate whenever the green light comes on. Both first-order and second-order classical conditioning are subject to extinction. Another limitation of classical conditioning involves the length of the delay between the presentation of the CS and the presentation of the UCS (in general, one second is optimal). Operant (instrumental) conditioning (aka “consequence learning”) Operant conditioning concerns the effect certain kinds of consequences have on the frequency of behavior. A consequence that increases the frequency of a behavior is called a reinforcement. A consequence that decreases the frequency of a behavior is called a punishment. Whether a consequence is reinforcing or punishing varies according to the person and the situation. There are two basic strategies for increasing the frequency of a behavior: positive reinforcement and negative reinforcement. There are two basic strategies for decreasing the frequency of a behavior: extinction and punishment. Edward L. Thorndike (1874-1949) Edward L. Thorndike graduated from Wesleyan University in 1895, and received his Ph.D. from Columbia University in 1898 He was appointed as an instructor in genetic psychology at Teachers College, Columbia University, in 1899, and served there until 1940. He devised methods to measure children’s intelligence and their ability to learn. He also conducted studies in animal psychology and the psychology of learning. His law of effect addressed the phenomenon we now call reinforcement. Thorndike’s books include Educational Psychology (1903), Mental and Social Measurements (1904), Animal Intelligence (1911), A Teacher’s Word Book (1921), Your City (1939), and Human Nature and the Social Order (1940). The Skinner box Operant (instrumental) conditioning (aka “consequence learning”) Operant conditioning concerns the effect certain kinds of consequences have on the frequency of behavior. A consequence that increases the frequency of a behavior is called a reinforcement. A consequence that decreases the frequency of a behavior is called a punishment. Whether a consequence is reinforcing or punishing varies according to the person and the situation. There are two basic strategies for increasing the frequency of a behavior: positive reinforcement and negative reinforcement. There are two basic strategies for decreasing the frequency of a behavior: extinction and punishment. Operant conditioning procedures Procedure Purpose Application Positive reinforcement Increase behavior Give reward following behavior Negative reinforcement Increase behavior Remove aversive stimulus following behavior Extinction Decrease behavior Do not reward behavior Punishment Decrease behavior Give aversive stimulus following behavior or take away positive stimulus Problems with the use of punishment Punishment does not teach what behavior is appropriate. It only teaches what behavior is inappropriate. To be effective, punishment must be delivered immediately and consistently. Punishment can have the negative side effect of inhibiting not only the undesirable behavior but also desirable behavior that is associated with it. Punishment can result in the person who is punished coming to fear the person who administers the punishment. Punishment may also serve as a behavior that is later modeled by the person being punished. Punishment can create strong negative emotions that can interfere with learning the desired response. For all of these reasons, punishment should be used sparingly and only when other operant conditioning procedures either cannot be used or will not work. Other important operant conditioning concepts Shaping: reinforcing successive approximations of the desired behavior until the complete response is well established Generalization: displaying the response to stimulus situations that resemble the one in which the original response was acquired Discrimination: selectively reinforcing the response to help ensure that it will only occur in the presence of the original stimulus and not ones that might resemble it Social Learning Theory Julian B. Rotter (1916- ) Julian Rotter first learned about psychology in the Avenue J Library in Brooklyn, where he spent much of his childhood and adolescence. Believing that he couldn’t earn a living as a psychologist, he majored in chemistry at Brooklyn College. While still in college, he discovered that Alfred Adler was teaching at the Long Island School of Medicine and began associating with Adler and his colleagues. He then became a psychology major at the University of Iowa and received a Ph.D. in clinical psychology at the University of Indiana. After serving as a psychologist in the Army during World War II, he later taught at the Ohio State University and at the University of Connecticut. Key concepts in Rotter’s Social Learning Theory: perceptions, expectancies, and values Behavior potential (BP): the likelihood of a given behavior occurring in a particular situation Expectancy: the perceived likelihood that a given behavior will result in a particular outcome – Generalized expectancies: beliefs about how often our actions typically lead to reinforcements and punishments – Locus of control: generalized perceptions about the degree to which one’s outcomes are determined by internal versus external factors Reinforcement value: the degree to which we prefer one reinforcer over another Rotter’s basic formula for predicting behavior Behavior potential (BP) = Expectancy (E) X Reinforcement Value (RV) If either the expectancy or the reinforcement value is zero, then the behavior potential will be zero. Examples of calculating the behavior potentials in an insult situation Option Possible outcome Expectancy Value Behavior potential Ask for apology Apology High High High Insult back Laughter Low High Average Yell at insulter Ugly scene High Low Average Leave the party Feel foolish Average Low Low Social-Cognitive Theory Albert Bandura (1925 Albert Bandura was raised in Alberta, Canada and received his bachelor’s degree at the University of British Columbia. He completed his Ph.D. at the University of Iowa, where he was influenced by the prominent learning theorist Kenneth Spence. After a year of clinical internship in Wichita, he accepted a position at Stanford University. His career has been spent building bridges between traditional learning theory, cognitive personality theory, and clinical psychology. ) Bandura’s reciprocal determinism model Behavior External factors (Rewards, punishments) Internal factors (Beliefs, thoughts, expectations) Features of learning and cognition that are (relatively) unique to humans The use of symbols and other cognitive representational structures to “re-create” the outside world within our own minds The resulting abilities to imagine alternative courses of action and conduct mental stimulations to “project” what their outcomes are likely to be The capacity for self-regulation through the application of self-reward and self-punishment, even in the face of strong external rewards and punishment The capacity for vicarious or observational learning Observational learning The distinction between learning and performance is important: not every behavior that is learned gets performed Much of our learning occurs vicariously, through our observation of other people’s actions, and the consequences of those actions. We are more likely to imitate a behavior we have seen other people display if the outcome of their behavior was a reward, rather than a punishment. In a study by Bandura (1965), nursery school children were more likely to model the aggressive behaviors of an adult who they observed in a film segment if they saw the model get rewarded, rather than punished, for his aggressive behavior. Mean number of aggressive responses performed (Bandura, 1965) 4 3.5 3 2.5 Boys 2 Girls 1.5 1 0.5 0 Model rewarded Model punished No consequences Diagram of Little Albert’s classical conditioning: A conditioned phobia Loud noise Fear responses (Unconditioned stimulus) (Unconditioned response) White rat (Conditioned stimulus) Fear response (Conditioned response) Application: behavior modification Classical conditioning applications – Systematic desensitization (for example, snake phobia) – Aversion therapy (for example, alcoholism) Operant conditioning applications – Changing behavior by changing contingencies • Reward • Punishment • Extinction through nonreinforcement – Token economy – Biofeedback Application: self-efficacy therapy The difference between an outcome expectancy and an efficacy expectancy. Four sources of efficacy expectancies: – – – – Enactive mastery experiences Vicarious experiences Verbal persuasion (“coaching”) Physiological and affective states Guided mastery as a step-by-step approach to achieve enactive mastery experiences Problems that have been addressed through the application of selfefficacy beliefs include traumatic stress disorder, test anxiety, phobias, and bereavement. Behavioral observation methods Direct observation – Direct observation in the field, in the laboratory, or in the clinic – Analogue behavioral observation (for example, staging a dance for clients who are being treated for shyness) – Role-play – The importance of reliable observations Self-monitoring – The importance of consistent and objective self-observation – The therapeutic value of self-observation Observation by others (parents, teachers, nurses, etc.) Strengths and criticisms of the behavioral / social learning approach Strengths – The approach has a solid foundation in empirical research with humans and with infrahuman species. – It led to the development of useful therapeutic procedures involving behavior modification. These intervention procedures, which are relatively quick, inexpensive, and easily administered, assess baseline levels of behavior and establish objective criteria for behavior change. – The social-cognitive extensions of the behavioral approach have greatly expanded the range of phenomena that can be addressed. Criticisms – The approach gives inadequate attention to the role of heredity. – Not all responses can be successfully conditioned (e.g., fear of food). – Rewarding intrinsically-motivated behaviors can sometimes reduce their frequency of occurrence. – Reducing the problems of therapy patients to observable behaviors may fail to address the underlying problem in certain kinds of disorders.