Operant Conditioning
Operant Conditioning
Operant Conditioning
Historical Background
-Edward Thorndike: His work on the “Law of Effect” laid the groundwork for
operant conditioning.
> The Core Principle: Thorndike observed that behaviors followed by
satisfying consequences are more likely to be repeated, while those followed
by unpleasant consequences are less likely to occur again. He termed this
the “Law of Effect”.
Ex: if you score a basket in basketball and the coach gives you a high-five or
praise, you feel good and want to score more baskets to get more praise.
Ex: In practice, every time you miss a shot, you have to do extra push-ups.
But if you make the shot, no push-ups! So, you want to make more shots to
avoid doing push-ups.
Sport example: If a player is late to practice, the coach might take away their
chance to play in the game. The player doesn’t want to miss the game, so
next time they’ll show up on time.
Ex:Imagine a kid who gets candy every time they clean their room.
They clean their room regularly because they want the candy.
But if the candy stops coming when they clean, they will clean their room
less and less, and eventually, they might stop cleaning altogether.
Imagine you’re teaching a dog a trick. If you give it a treat every time it
does the trick (reinforcement), it’ll learn quickly. But if you stop giving treats,
the dog will eventually stop doing the trick because there’s no reward
anymore. That’s extinction.
In school, if a student acts up to get attention, and the teacher stops giving
them attention, the student will likely stop acting up because it’s no longer
working. That’s extinction too. It’s about removing the reward to make a
behavior disappear. It’s important to be consistent; if you sometimes give
the reward, it won’t work as well.
• Schedules of Reinforcement: These are rules that determine when and how
often reinforcement is delivered. Different schedules have different effects on
learning and behavior.
-Partial reinforcement means you get a reward sometimes, but not every
time you do something.
Ex: Imagine you only get a candy sometimes when you clean your room. You
won’t know exactly when you’ll get it, so you keep cleaning, hoping for a
reward.
>fixed-ratio schedule means you get a reward after doing something a set
number of times.
Ex: If you get a candy after every 3rd time you clean your room, that’s a fixed
ratio. No matter what, after you clean 3 times, you get the reward.
>Variable-ratio schedule means you get a reward, but you don’t know
exactly when—it happens after a random number of times you do something.
Ex: you clean your room, but you might get a candy after the 2 nd, 5th, or 10th
time. You never know when the next reward is coming, so you keep cleaning,
hoping the next time will get you the candy.
>Fixed-interval schedule means you get a reward after a set amount of time,
as long as you do the behavior during that time.
Ex: Imagine you can get a treat for cleaning your room, but only once every
week. Even if you clean your room a lot during the week, you’ll only get the
treat after that week is over.
Ex: Think about how you wait for the bus. Sometimes the bus comes after 5
minutes, sometimes after 10 minutes, and sometimes after 15 minutes.
•The Skinner Box: The Skinner box typically contained a lever or key that an
animal could press or peck. When the animal performed the desired
behavior, it received a reward, such as food or water. Skinner used this
apparatus to study how different schedules of reinforcement affected
learning and behavior.
Ex:
Experiments:
Setup:
Procedure:
Skinner would provide a food pellet as a reward whenever the rat pressed
the lever.
Findings:
Learning: The rat quickly learned to press the lever to get food,
demonstrating the principle of positive reinforcement—the reward (food)
encouraged the behavior (lever pressing).
Fixed-Ratio Schedule: The rat had to press the lever a specific number of
times to get food, which resulted in a burst of responses as the reward was
approached.
Setup:
Procedure:
The pigeons would peck at the disk, and Skinner would dispense food as a
reward for pecking.
Findings:
Learning: Like the rats, the pigeons learned that pecking the disk led to food.
Summary:
Education
Therapy:
Parenting:
Workplace:
Organizations use operant conditioning to motivate employees and improve
productivity. Performance-based bonuses, recognition programs, and
consequences for poor performance are all examples of operant conditioning
principles in the workplace.
- Oversimplification:
- Ethical Concerns:
Ethical concerns about punishment in operant conditioning focus on its
potential to cause fear, promote aggression, and lead to other negative
consequences. Critics argue that these outcomes can be harmful,
highlighting the need for careful consideration when using punishment
as a behavior modification tool.
- Individual Differences:
Individual differences highlight that people do not respond the same
way to reinforcement and punishment. What reinforces or discourages
one person may not have the same effect on another due to personal
preferences, experiences, and cultural influences.
- Complexity of Behavior:
Conclusion
Operant conditioning is a useful way to understand how behaviors are
learned and changed. Its principles can be applied in many areas,
including education, therapy, parenting, and animal training. However,
it’s important to recognize its limitations and use it carefully.
Understanding the complexity of human behavior and considering
ethical issues are crucial when applying these principles to ensure
positive and effective outcomes.