Chapter 8 Learning James Henderson Robert Reed Shalom Aziague Denise Delagarza Seth Ewing-Hendrick Classical Conditioning Classical conditioning More than 200 years ago, philosophers such as John Locke and David Hume echoed Aristotle's conclusion from 2000 years earlier. We learn by association our minds naturally connect events that occur in sequence. Example: If you associate a sound with a frightening consequence, then your fear may be aroused by the sound itself Cont.. By linking two events that occur close together, both the sea snail and the seal exhibit associate learning. Conditioning is the process of learning association In classical conditioning, we learn to associate two stimuli and thus to anticipate events. We learn that a flash of lightning signals an impending crack of thunder ,and so we start to brace ourselves when lightning strikes nearby In operant conditioning, we learn to associate a response (our behavior) and its consequences and thus to repeat acts followed by good results and avoid acts followed by bad results. Classical conditioning Ivan Pavlov performed experiments that are considered classical conditioning Another important psychologist that dealt with classical conditioning was John B. Watson. John’s idea was that psychology should study how organisms respond to stimuli in their environments. He thought that psychology should be an objective science based on observable behavior which is called behaviorism Pavlov experiment After studying salivary secretion in dogs ,he knew that when he put food in a dog’s mouth, the animal would invariably salivate The repeated behavior could make the dog salivate with the mere sight of the food, the smell of the food or even the person who brought it Because salivation in response to food in the mouth was unlearned, Pavlov called it an unconditional response (UR). Food in the mouth automatically, unconditionally triggers a dog’s salivary reflex. Thus Pavlov called the food stimulus and unconditioned stimulus (US). One translation of Pavlov therefore calls the salivation the conditional reflex. Today we call this learned response the conditional response (CR). The permissibly irrelevant tone stimulus that now triggered the conditional salivation we call the conditional stimulus (CS). Acquisition(Initial learning) To understand acquisition of the stimulus –response relationship, Pavlov and his associates first had to confront the question of timing. Although it’s not likely for conditioning to occur, it could occur when the CS follow the US. This finding fits the presumption that classical conditioning is biologically adaptive. It helps organisms prepare for good or bad events. Michael Domjan Showed how the CS signal are important biological event by conditioning the sexual arousal of male Japanese quail Ex: Just before presenting an approachable female, the researchers turned on a red light, over time ,the red light announced the arrival of the female quail, which caused the male quail to because sexually aroused and copulate with her more quickly when she arrived. Moreover, the quail grew to enjoy the red-light district of the cage. Exposure to the sexually conditioned stimuli also caused them to release more semen and sperm. The quail’s capacity for classical conditioning gives it a reproductive edge. Extinction and spontaneous recovery According to Pavlov experiment, when he sounded the tone again and again without presenting food, the dogs salivated less and less. The declining salivation illustrates extinction. If allowed several hours to elapse before sounding the tone again, the salivation to the tone would reappear spontaneously Generalization Pavlov and his students noticed that a dog conditioned to the sound of one tone also responded somewhat to the sound of a different tone never paired with food. The tendency to respond to stimuli similar to the CS is called generalization Example: generalization can be adaptive, as when toddlers taught to fear moving cars in the street respond similarly to trucks and motorcycles. Discrimination Pavlov dogs also learned to respond to the sound of a particular tone and not to other tones. Like dog generalization, discrimination has survival value. Slightly different stimuli are at time followed by vastly different consequences Extending Pavlov’s understanding Pavlov’s and Watson’s Disdain for “mentalist” concepts such as consciousness has given way to a growing realization that they underestimated the importance of cognitive processes(thoughts, perceptions , expectations)and biological constraints on an organism’s learning capacity Cognitive processes Robert Rescorla and Allan Wagner argued that when two significant events occur close together in time, an animal learns the predictability of the second event. An example of this is if a shock is always precede by a tone, and then sometimes also by a light that accompanies the tone, a rat will react with fear to the tone but not to the light. Although the light is followed by the shock ,it adds no new information; the tone better predicts the impending shock. That principle helps explain why classical conditioning treatments that ignore cognition often have limited success. Biological predispositions Ever since Darwin, scientists have assumed that all animals share a common evolutionary history and resulting commonalities in their makeup and functionality. Pavlov and Watson believed the basic laws of learning were essentially similar in all animals, so it should make a little difference between pigeons and humans. More than the early behaviorists realized, an animals capacity for conditioning is constrained by its biology. Each species' predispositions prepare it to learn the associations that enhance its survival. Operant Conditioning Operant Conditioning Operant conditioning explains and trains behaviors such as an elephant walking on its hind legs or a child saying please. Classical and operant conditioning include acquisition, extinction, spontaneous recovery, generalization, and discrimination. Operant behavior produces rewarding or punishing stimuli (consequences). Skinner’s Experiments B. F. Skinner was a college English major and an aspiring writer who, seeking a new direction, entered graduate school in psychology. He went on to become modern behaviorism’s most influential and controversial figure. His work elaborated a simple fact of life that psychologist Edward L. Thorndike called the law of effect, which stated rewarded behavior is likely to recur. He used the law of effect as a starting point and developed a “behavioral technology” that revealed principles of behavior control. These principles allowed him to teach pigeons unpigeon like behavior such as walking in a figure 8, playing Ping-Pong, and keeping a missile on course by pecking at a target on a screen. Skinner’s Experiments cont. Skinner designed an operant chamber, known as a Skinner Box which he used for his studies with rats and eventually pigeons. The box has a bar or key that an animal presses or pecks at to release a reward of food or water, and a device that records these responses. Shaping Behavior Uses reinforcers, such as food, to guide an animal’s actions toward a desired behavior. Makes the animal get closer and closer to where you want and rewards them. Sometimes we as humans reward annoying behaviors unintentionally. Types of Reinforcers Positive reinforcement strengthens a response by presenting a typically pleasurable stimulus after a response. Positive reinforcers include: praise, attention, hugs, food, etc. Negative reinforcement strengthens a response by reducing or removing an aversive (undesirable) stimulus. Negative reinforcers include: taking aspirin to relieve a headache and pushing the snooze button to silence an annoying alarm clock. Primary and Conditioned Reinforcers Primary reinforcers, such as getting food when hungry or being relieved of electric shock, are innately satisfying. Conditioned reinforcers, also known as secondary reinforcers, are learned. They get their power through their association with primary reinforcers. Such as if a rat in a Skinner box learns that a light reliably signals that food is coming, the rat will work to turn on the light. Immediate and Delayed Reinforcers Rats do not respond well to delayed reinforcers. The rat in a Skinner box will engage in “unwanted” behavior (scratching, sniffing, and moving around) before performing the “wanted” behavior to receive the reward. If the rat presses the bar, but if the researcher is distracted and delays the reinforcer for longer than 30 seconds, the rat will not learn to press the bar. Unlike rats, humans do respond to reinforcers that are greatly delayed: the paycheck at the end of the week, the good grade at the end of the semester. To function effectively we must learn to postpone immediate rewards for greater long-term reward. Reinforcement Schedules Many of the previous examples assume continuous reinforcement: The desired response is reinforced every time it occurs. Under these conditions, learning occurs rapidly, but so does extinction. Real life does not often provide continuous reinforcement, so why do people do the things they do? They do it because their efforts have occasionally been rewarded. This is called partial (intermittent) reinforcement schedule, in which the responses are sometimes reinforced, sometimes not. Initial learning is slower with intermittent reinforcement, but it produces greater persistencegreater resistance to extinction. Reinforcement Schedules cont. Fixed-ratio schedules reinforce behavior after a set number of responses much like people paid on a piecework basis. Variable-ratio schedules provide reinforcers after an unpredictable number of responses. This is what gamblers experience and why it is so hard to extinguish. Fixed-interval schedules reinforce the first response after a fixed time period. Like people checking more frequently for the mail as the delivery time approaches. Variable-interval schedules reinforce the first response after varying time intervals. Punishment The effect of punishment is opposite to that of reinforcement. Punishment decreases a behavior, thus a punisher is any consequence that decreases the frequency of a preceding behavior, usually by administering an undesirable consequence or withdrawing a desirable one. Physical punishment may increase aggressiveness by demonstrating that aggression is a way to cope with problems. This helps explain why so many aggressive delinquents and abusive parents come from abusive families. Even though punishment suppresses unwanted behavior, it often does not guide one toward more desirable behavior. Punishment tells you what not to do; reinforcement tells you what to do. Latent Learning Rats exploring a maze, with no obvious reward, are like people sightseeing in a new town. The rats seem to develop a cognitive map, a mental representation of the maze. When an experimenter then places a reward in the maze’s goal box, the rats very quickly perform as well as rats that have been reinforced with food for running the maze. During their explorations, the rats seemingly experience latent learning, learning that becomes apparent only when there is some incentive to demonstrate it. Intrinsic Motivation Excessive rewards can undermine intrinsic motivation, the desire to perform a behavior effectively and for its own sake. Intrinsically motivated people work and play in search of enjoyment, interest, selfexpression, or challenge. Extrinsic motivation is the desire to behave in certain ways to receive external rewards or avoid threatened punishment. Skinner’s Legacy Applications of operant conditioning can be seen at school, in sports, at work, and at home. Contrasting Classical and Operant Conditioning Both classical and operant conditioning are forms of associative learning Classical conditioning is when an organism associates different stimuli that it has no control over and responds to it. Through operant conditioning, an organism associates operant behaviors with their consequences. Learning by Observation Not all learning needs to occur through direct experience Observational learning is when we observe and imitate others This learning type also applies to other species Cultural elements gained through imitation are called memes Mirror neurons provide a neural basis for observational learning Mirror neurons also serve language, help children learn how to mime lip and tongue movements when forming new words. They also allow us to infer another’s mental state Bandura’s Experiments While a child is in the room, an adult beats a Bobo boll for a long period of time. Later when the kid is frustrated and left alone, he also attacks the Bobo doll using the same words the adult used. Watching the doll be beaten lowered the child’s inhibitions What determines what we imitate are reinforcements and punishments By looking we can determine a behaviors consequences We are more likely to imitate those we see as similar to ourselves as successful or as admirable Applications of Observable Learning Anti-social models may have anti-social effects Some learned things are that physical intimidation is effective and free and easy sex is pleasurable with no consequences. Also explains how abusive parents have aggressive kids and wife-battering fathers had parents who battered others. Lessons learned as children are not easily forgotten Pro-social models can have positive effects on others Parents re also powerful models Those exposed to a hypocrite model tend to become hypocrites themselves Television is a huge source of observational learning seeing as most children when they are 18 watch more television than they spend in school Applications of Observable Learning cont. T.V. allows for spread of culture between different countries Television portrays the real world poorly Before finishing elementary school, the average child has seen 8000 television murders and 100000 other acts of violence Violence is in 6 in every 10 programs, 78% goes unpunished, 58% did not show victims pain, and nearly half is seen as justified. The more media violence a child sees, the more often they get into fights The more violence watched, the more at risk they are for aggression and crime Homicide rates doubled between 1957 and 1974 when television was introduced Applications of Observational Learning cont. While there is a correlation, it does not prove causation There is a possibility that violent programs reflect violent behavior, not that it causes it The scientific community believes media violence is a cause of violence, especially if caused by an attractive person who goes unpunished and causes no visible pain or harm. Prolonged exposure to violence desensitizes viewers