This document discusses classical conditioning as discovered by Ivan Pavlov. It defines key terms like unconditioned stimulus, unconditioned response, conditioned stimulus, and conditioned response. Through Pavlov's famous experiment pairing food with a metronome for dogs, he found that a neutral stimulus could become associated with an unconditioned response and elicit a conditioned response through repeated pairing with an unconditioned stimulus in close temporal proximity. The summary discusses how classical conditioning is a simple form of learning important principles like temporal contiguity between the conditioned and unconditioned stimulus.
This document discusses classical conditioning as discovered by Ivan Pavlov. It defines key terms like unconditioned stimulus, unconditioned response, conditioned stimulus, and conditioned response. Through Pavlov's famous experiment pairing food with a metronome for dogs, he found that a neutral stimulus could become associated with an unconditioned response and elicit a conditioned response through repeated pairing with an unconditioned stimulus in close temporal proximity. The summary discusses how classical conditioning is a simple form of learning important principles like temporal contiguity between the conditioned and unconditioned stimulus.
This document discusses classical conditioning as discovered by Ivan Pavlov. It defines key terms like unconditioned stimulus, unconditioned response, conditioned stimulus, and conditioned response. Through Pavlov's famous experiment pairing food with a metronome for dogs, he found that a neutral stimulus could become associated with an unconditioned response and elicit a conditioned response through repeated pairing with an unconditioned stimulus in close temporal proximity. The summary discusses how classical conditioning is a simple form of learning important principles like temporal contiguity between the conditioned and unconditioned stimulus.
This document discusses classical conditioning as discovered by Ivan Pavlov. It defines key terms like unconditioned stimulus, unconditioned response, conditioned stimulus, and conditioned response. Through Pavlov's famous experiment pairing food with a metronome for dogs, he found that a neutral stimulus could become associated with an unconditioned response and elicit a conditioned response through repeated pairing with an unconditioned stimulus in close temporal proximity. The summary discusses how classical conditioning is a simple form of learning important principles like temporal contiguity between the conditioned and unconditioned stimulus.
Download as PPTX, PDF, TXT or read online from Scribd
Download as pptx, pdf, or txt
You are on page 1of 45
By: Pooja Upadhyay
“Learning is any relatively permanent change in
behavior brought about by experience or practice.” The “relatively permanent” part of the definition refers to the fact that when people learn anything, some part of their brain is physically changed to record what they’ve learned. This is actually a process of memory, for without the ability to remember what happens, people cannot learn anything. Although there is no conclusive proof as yet, research suggests that once people learn something, it may be present somewhere in memory in physical form They may be unable to “get” to it, but it’s there. Not all change is accomplished through learning. Changes like an increase in height or the size of the brain are another kind of change, controlled by a genetic blueprint, generally referred to as maturation, that happens due to biology and not learning. For example, practice alone will not allow a child to walk. Children learn to walk because their nervous systems, muscle strength, and sense of balance have reached the point where walking is physically possible for them—all factors controlled by maturation. Once that maturational readiness has been reached, then practice and experience play their important part. It is important to distinguish between learning and performance. Performance is simply what a person or animal does, but that’s only one ingredient to the whole process. There are other several learned and unlearned factors involved that affect what we do. These factors interact with each other in complex ways, for example, learned behaviour depends on motivation, a rat that has learned the maze does not performs well unless its hungry. Hence in this case learning does not result in performance unless it is brought out through motivation, so psychologists have to be very careful while understanding the various factors that might affect performance. One of the important conditions for learning is arousal in the right amount. One should not be too less aroused, otherwise the learning would not take place. The person should also not be too much aroused, most forms of learning are retarded. Up to a certain point, the higher the arousal, the higher the arousal level, the better the learning. CLASSICAL CONDITIONING OPERANT CONDITIONING OBSERVATIONAL LEARNING Itwas a Russian physiologist (a person who studies the workings of the body) named Ivan Pavlov (1849–1936) who pioneered the empirical study of the basic principles of a particular kind of learning. Studying the digestive system in his dogs, Pavlov had built a device that would accurately measure the amount of saliva produced by the dogs when they were fed a measured amount of food. Normally, when food is placed in the mouth of any animal, the salivary glands automatically start releasing saliva to help with chewing and digestion. This is a normal reflex—an unlearned, involuntary response that is not under personal control or choice— one of many that occur in both animals and humans. The food causes a particular reaction, the salivation. A stimulus can be defined as any object, event, or experience that causes a response, the reaction of an organism. In the case of Pavlov’s dogs, the food is the stimulus and salivation is the response. Pavlov soon discovered that his dogs began salivating when they weren’t supposed to be salivating. Some dogs would start salivating when they saw the lab assistant bringing their food, others when they heard the clatter of the food bowl from the kitchen, and still others when it was the time of day they were usually fed. Switching his focus, Pavlov spent the rest of his career studying what eventually he termed classical conditioning, learning to elicit an involuntary, reflex- like, response to a stimulus other than the original, natural stimulus that normally produces the response Unconditioned Stimulus: • The original, naturally occurring stimulus is called the unconditioned stimulus (UCS). • The term unconditioned means “unlearned.” • This is the stimulus that ordinarily leads to the involuntary response. • In the case of Pavlov’s dogs, the food is the unconditioned stimulus. Unconditioned Response : • automatic and involuntary response to the unconditioned stimulus is called the unconditioned response (UCR) for much the same reason. • It is unlearned and occurs because of genetic “wiring” in the nervous system. • For example, in Pavlov’s experiment, the salivation to the food is the UCR (unconditioned response). Conditioned Stimulus: • Pavlov determined that almost any kind of stimulus could become associated with the unconditioned stimulus (UCS) if it is paired with the UCS often enough. • In his original study, the sight of the food dish itself became a stimulus for salivation before the food was given to the dogs. • Every time they got food (to which they automatically salivated), they saw the dish. • At this point, the dish was a neutral stimulus (NS) because it had no effect on salivation. • After being paired with the food so many times, the dish came to produce a salivation response, although a somewhat weaker one, as did the food itself. When a previously neutral stimulus, through repeated pairing with the unconditioned stimulus, begins to cause the same kind of involuntary response, learning has occurred. • The previously neutral stimulus can now be called a conditioned stimulus (CS). (Conditioned means “learned,” and, as mentioned earlier, unconditioned means “unlearned.”) Conditioned Response: • The response that is given to the CS (conditioned stimulus) is not usually quite as strong as the original unconditioned response (UCR), but it is essentially the same response. • However, because it comes as a learned response to the conditioned stimulus (CS), it is called the conditioned response (CR). Pavlov did a classic experiment in which he paired the ticking sound of a metronome (a simple device that produces a rhythmic ticking sound) with the presentation of food to see if the dogs would eventually salivate at the sound of the metronome. Since the metronome’s ticking did not normally produce salivation, it was a neutral stimulus (NS) before any conditioning took place. The repeated pairing of a NS and the UCS (unconditioned stimulus) is usually called acquisition, because the organism is in the process of acquiring learning Notice that the responses, CR (conditioned response) and UCR (unconditioned response), are very similar—salivation. However, they differ not only in strength but also in the stimulus to which they are the response. An unconditioned stimulus (UCS) is always followed by an unconditioned response (UCR), and a conditioned stimulus (CS) is always followed by a conditioned response (CR). Classical conditioning is actually one of the simplest forms of learning. It’s so simple that it happens to people all the time without them even being aware of it. • Does your mouth water when you merely see an advertisement for your favorite food on television? • Do you feel anxious every time you hear the high-pitched whine of the dentist’s drill? Over the course of many visits to the dentist, for example, the body comes to associate that sound (CS) with the anxiety or fear (UCR) the person has felt while receiving a painful dental treatment (UCS), and so the sound produces a feeling of anxiety (CR) whether that person is in the chair or just in the outer waiting area. 1. The CS must come before the UCS. If Pavlov sounded the metronome just after he gave the dogs the food, they did not become conditioned. 2. The CS and UCS must come very close together in time—ideally, no more than 5 seconds apart. When Pavlov tried to stretch the time between the potential CS and the UCS to several minutes, no association or link between the two was made. Too much could happen in the longer interval of time to interfere with conditioning. Recent studies have found that the interstimulus interval can vary depending on the nature of the conditioning task and even the organism being conditioned. In these studies, shorter ISIs (less than 500 milliseconds) have been found to be ideal for conditioning. 3. The neutral stimulus must be paired with the UCS several times, often many times, before conditioning can take place. 4. The CS is usually some stimulus that is distinctive or stands out from other competing stimuli. The metronome, for example, was a sound that was not normally present in the laboratory and, therefore, distinct . The tendency to respond to a stimulus that is similar to the original conditioned stimulus is called stimulus generalization. For example, a person who reacts with anxiety to the sound of a dentist’s drill might react with some slight anxiety to a similar-sounding machine, such as an electric coffee grinder. Stimulus discrimination occurs when an organism learns to respond to different stimuli in different ways. For example, although the sound of the coffee grinder might produce a little anxiety in the dental-drill-hating person, after a few uses that sound will no longer produce anxiety because it isn’t associated with dental pain. What would have happened if Pavlov had stopped giving the dogs food after the real CS? Pavlov did try just that, and the dogs gradually stopped salivating to the sound of the ticking. When the metronome’s ticking (CS or conditioned stimulus) was repeatedly presented in the absence of the UCS (unconditioned stimulus or food, in this case), the salivation (CR or conditioned response) “died out” in a process called extinction. After extinguishing the conditioned salivation response in his dogs, Pavlov waited a few weeks, putting the conditioned stimulus (i.e., the metronome) away. There were no more training sessions and the dogs were not exposed to the metronome’s ticking in that time at all. But when Pavlov took the metronome back out and set it ticking, the dogs all began to salivate, although it was a fairly weak response and didn’t last very long. This brief recovery of the conditioned response proves that the CR is “still in there” somewhere. As time passes, this inhibition weakens, especially if the original conditioned stimulus has not been present for a while. In spontaneous recovery the conditioned response can briefly reappear when the original CS returns, although the response is usually weak and short-lived. Edward L. Thorndike was one of the first researchers to explore and attempt to outline the laws of learning voluntary responses, although the field was not yet called operant conditioning. Thorndike placed a hungry cat inside a “puzzle box” from which the only escape was to press a lever located on the floor of the box. Thorndike placed a dish of food outside the box, so the hungry cat is highly motivated to get out. Thorndike observed that the cat would move around the box, pushing and rubbing up against the walls in an effort to escape. Eventually, the cat would accidentally push the lever, opening the door. Upon escaping, the cat was fed from a dish placed just outside the box. The lever is the stimulus, the pushing of the lever is the response, and the consequence is both escape (good) and food (even better). The cat did not learn to push the lever and escape right away. After a number of trials (and many errors) in a box like this one, the cat took less and less time to push the lever that would open the door. It’s important not to assume that the cat had “figured out” the connection between the lever and freedom. Thorndike kept moving the lever to a different position, and the cat had to learn the whole process over again. The cat would simply continue to rub and push in the same general area that led to food and freedom the last time, each time getting out and fed a little more quickly. Based on this research, Thorndike developed the law of effect: If an action is followed by a pleasurable consequence, it will tend to be repeated. If an action is followed by an unpleasant consequence, it will tend not to be repeated. This is the basic principle behind learning voluntary behavior. In the case of the cat in the box, pushing the lever was followed by a pleasurable consequence (getting out and getting fed), so pushing the lever became a repeated response. B. F. Skinner was the behaviorist who assumed leadership of the field after John Watson. He was even more determined than Watson that psychologists should study only measurable, observable behavior. Skinner found in the work of Thorndike a way to explain all behavior as the product of learning. He even gave the learning of voluntary behavior a special name: operant conditioning. Voluntary behavior is what people and animals do to operate in the world. Classical conditioning is the kind of learning that occurs with automatic, involuntary behavior. The kind of learning that applies to voluntary behavior is called operant conditioning, which is both different from and similar to classical conditioning. Operant conditioning consists of learning to perform some act that leads to reward or punishment. An organism in the operant conditioning emits response like wandering around, pushing, looking at things rather than giving a specific elicited response as in classical conditioning. CLASSICAL CONDITIONING OPERANT CONDITIONING
1. Stimulus is a specific event, Stimulus is not specific, it is a
such as flash of a light, or the longer lasting situation which has sound of a tone, which is briefly several features, only one or few presented. of which prove relevant to 2. The response, like the stimulus ,is a specific one. Moreover, it is learning. usually a reflex or an innate Responses are first varied, random reaction to a situation. movements in the stimulus 3. Reinforcement is always situation. presented following the Reinforcement does not depend conditioning situation, regardless on the response. If the subject does of what the person or animal the right thing then only it is does. reinforced, otherwise not. Operant conditioning can be illustrated by understanding an experiment done by Skinner on his laboratory rats. There was a chamber that could be outfitted in many ways --with two or more levers, two or more lights, a feeding place into which pellets can be dropped, a drinking place for water, and a metal grid floor for applying electric shock. The chamber consists of a box with lever protruding from a wall and a food cup below it for rewarding the animal with food. Attached to the lever is a electrical circuit, a device that makes a recording on paper each time he rat pushes the lever. It is called a cumulative recorder, because one response moves the pen one unit, another response another unit and so on.. As the responses are cumulative they add up. Since the paper moves on a constant speed, a steep line on the record means that the rat is making responses in quick succession. A relatively flat line means that it is making very few responses. Suppose now a hungry rat is put in the operant chamber hooked up to the cumulative recorder. Since the box is unfamiliar to the rat, it would provoke fear in it. But these signs fade soon as the box becomes more and more familiar, and the rat starts to explore. It does many things like sniffing the wall, paws at the floor and walls, stands on its hind legs, and runs along the floor. Eventually it depresses the lever, perhaps by leaning on it, or by bumping its head, or by grasping it by his paws. When the lever is pressed a pallet of food is released and falls into the food cup. This is the rat’s first correct response and first reward in the chamber. The period between the time the rat was placed in the box and time it made the first rewarded response was 15 minutes. And there was another minute or so before the rat noticed the food and ate it. The rat didn’t learn much from the first experience, but the food palette that the rat ate because it was hungry aroused it and caused it to explore with more vigor. After a few incidences of the same, the rat got an idea or he had become conditioned. Responses started coming rapidly, so it pushed the lever and ate the pellet as fast as a rat can. The experiment illustrates the conditioning of an operant response. The rat’s operant behaviour, its repertoire of responses, at the beginning of the session had many components like sniffing, pawing, running, standing, and incidentally pressing the bar. But only one of the responses were rewarded. This response was the one it started learning after a few trials of pairing the response with the reward. Note carefully that the rat was required to emit the response itself , it was not a reflex-like response that was elicited, as it is in the case of classical conditioning. “What’s in it for me?” represents the concept of reinforcement, one of Skinner’s major contributions to behaviorism. The word itself means “to strengthen,” and Skinner defined reinforcement as anything that, when following a response, causes that response to be more likely to happen again. Typically, this means that reinforcement is a consequence that is in some way pleasurable to the organism, which relates back to Thorndike’s law of effect. The “pleasurable consequence” is what’s “in it” for the organism. Keep in mind that a pleasurable consequence might be something like getting food when hungry or a paycheck when you need money, but it might also mean avoiding a tiresome chore, like doing the dishes or taking out the garbage. I’ll do almost anything to get out of doing the dishes, myself! Primary Secondary
Gets its reinforcing
properties from being Fulfils basic needs associated with primary reinforcers
Food, candy bar money
POSITIVE REINFORCEMENT NEGATIVE REINFORCEMENT
the reinforcement of a response Following a response with the
by the addition or experience of removal or escape from a pleasurable consequence, something unpleasant will also such as a reward or a pat on the increase the likelihood of that back. response being repeated. The timing of reinforcement can make a tremendous difference in the speed at which learning occurs and the strength of the learned response. Consider the following scenario: Alicia’s mother agrees to give her a quarter every night she remembers to put her dirty clothes in the clothes hamper. Bianca’s mother agrees to give her a dollar at the end of the week, but only if she has put her clothes in the hamper every night. Alicia learns to put her clothes in the hamper more quickly than does Bianca because responses that are reinforced each time they occur are more easily and quickly learned. After a time, the mothers stop giving the girls the money. Which child is more likely to stop putting her clothes in the hamper? The answer might surprise you. It is more likely that Alicia, who has expected to get a reinforcer (the quarter) after every single response, will stop putting her clothes in the hamper. As soon as the reinforcers stop for her, the behavior is no longer reinforced and is likely to extinguish. In contrast, Bianca has expected to get a reinforcer only after seven correct responses. When the reinforcers stop for her, she might continue to put the clothes in the hamper for several more days or even another whole week, hoping that the reinforcer will eventually come anyway. Bianca may have learned more slowly than Alicia, but once she learned the connection between putting her clothes in the hamper and getting that dollar, she is less likely to stop doing it—even when her mother fails to give the dollar as expected. Bianca’s behavior illustrates the partial reinforcement effect. A response that is reinforced after some, but not all, correct responses will be more resistant to extinction than a response that receives continuous reinforcement (a reinforcer for each and every correct response). Although it may be easier to teach a new behavior using continuous reinforcement, partially reinforced behavior is not only more difficult to suppress but also more like real life. Imagine being paid for every hamburger you make or every report you turn in. In the real world, people tend to receive partial reinforcement rather than continuous reinforcement for their work. Ina fixed ratio schedule of reinforcement, the number of responses required to receive each reinforcer will always be the same number.
A variable ratio schedule of reinforcement is
one in which the number of responses changes from one trial to the next. An important characteristic of learners is the previous learning that they bring to a particular task. Fully grown learners never learn anything from scratch. Since they have had years of learning both in and out of school, any new learning builds on the old. This effect is called transfer of training, which can be either a help (positive transfer) or a hindrance (negative transfer). There are rules that control positive or negative transfer of learning: 1. Similarity of stimuli – positive transfer between old and new tasks favored by similarity of stimuli. It makes the learning of the task easier. 2. Similarity of responses – positive transfer is also favored by similarity of responses in two tasks. For example, tennis and badminton.