Operant Conditioning: Learn It 1—The Law of Effect

  • Define and give examples of operant conditioning
  • Understand the difference between reinforcement and punishment
  • Identify types of reinforcement and punishment

Operant Conditioning

In the previous section, you learned about classical conditioning, where an environmental stimulus automatically triggers a reflexive response. The learner is passive—responses happen without choice.
Now we turn to the second major form of associative learning: operant conditioning, sometimes called instrumental conditioning.

Operant conditioning focuses on how voluntary behaviors are shaped by their consequences.

operant conditioning

In operant conditioning, an organism learns to associate a behavior with a consequence. If the consequence is pleasant, the behavior becomes more likely to happen again. If the consequence is unpleasant, the behavior becomes less likely to occur.

For example, Spirit, a dolphin at the National Aquarium in Baltimore, does a flip in the air when her trainer blows a whistle. The consequence, or reward, is that she gets a fish.

How Operant Conditioning Differs from Classical Conditioning

There can sometimes be gray areas in differentiating between what counts as classical or operant conditioning. An important distinction is that classical conditioning emphasizes the passive learning process where the learner responds involuntarily in an automatic, reflexive manner, while operant conditioning is active and involves choice. In operant conditioning, the learner voluntarily chooses if they want to perform the behavior to get the reward.

For example, Pavlov’s dogs couldn’t help but drool when hearing the bell (classical), but they could have made a choice to “roll over” in anticipation of a treat (operant).

The table below also highlights how in classical conditioning, the stimulus comes before the response, while in operant conditioning, it follows the behavior.

Table 1. Classical and Operant Conditioning Compared
  Classical Conditioning Operant Conditioning
Conditioning approach

A neutral stimulus becomes associated with an unconditioned stimulus to produce a reflexive response.

A voluntary behavior is followed by reinforcement or punishment to influence whether it happens again.

Stimulus timing The stimulus occurs immediately before the response. The stimulus (either reinforcement or punishment) occurs soon after the response.

The Law of Effect: The Foundation of Operant Conditioning

Psychologist B. F. Skinner saw that classical conditioning is limited to existing behaviors that are reflexively elicited, and it doesn’t account for new behaviors such as riding a bike. He proposed a theory about how such behaviors come about.

Skinner believed that behavior is motivated by the consequences we receive for the behavior: the reinforcements and punishments. Skinner’s idea that learning is the result of consequences is based on the law of effect, which was first proposed by psychologist Edward Thorndike.

the law of effect

According to the law of effect, behaviors that are followed by consequences that are satisfying to the organism are more likely to be repeated, and behaviors that are followed by unpleasant consequences are less likely to be repeated (Thorndike, 1911).

 

Essentially, if an organism does something that brings about a desired result, the organism is more likely to do it again. If an organism does something that does not bring about a desired result, the organism is less likely to do it again. 

An example of the law of effect is in employment. One of the reasons (and often the main reason) we show up for work is because we get paid to do so. If we stop getting paid, we will likely stop showing up—even if we love our job.

Thorndike tested this principle using puzzle boxes with cats, observing how long it took them to escape after discovering a lever or latch. Over repeated trials, escape time decreased—evidence that rewarding outcomes strengthened behavior.

This video demonstrates how Thorndike experimented with cats to develop his ideas about learning.

The Skinner Box

Working with Thorndike’s law of effect as his foundation, Skinner began conducting scientific experiments on animals (mainly rats and pigeons) to determine how organisms learn through operant conditioning (Skinner, 1938). He placed these animals inside an operant conditioning chamber, which has come to be known as a “Skinner box” (Figure 1).

A Skinner box contains a lever (for rats) or disk (for pigeons) that the animal can press or peck for a food reward via the dispenser. Speakers and lights can be associated with certain behaviors. A recorder counts the number of responses made by the animal.

A photograph shows B.F. Skinner. An illustration shows a rat in a Skinner box: a chamber with a speaker, lights, a lever, and a food dispenser.
Figure 1. (a) B. F. Skinner developed operant conditioning for the systematic study of how behaviors are strengthened or weakened according to their consequences. (b) In a Skinner box, a rat presses a lever in an operant conditioning chamber to receive a food reward. (credit a: modification of work by “Silly rabbit”/Wikimedia Commons)
Watch the following clip to learn more about operant conditioning and to watch an interview with Skinner as he talks about conditioning pigeons.
You can view the transcript for “Operant conditioning” here (opens in a new window).