Understanding Learning Types
Understanding Learning Types
DEFINITION:
Learning is a relatively permanent change in behavior that results from experience. It is the acquisition of
information, knowledge, and skills.
The term "active learning" is often used to describe an interactive process, such as doing a hands-
on experiment to learn a concept rather than reading about it. But "passive learning" (reading a
text, listening to a lecture, watching a movie) is still learning, and can be effective.
Types of Learning:
1. Motor learning:
Most of our activities in our day-to-days life refer to motor activities.
The individual has to learn them in order to maintain his regular life,
for example walking, running, skating, driving, climbing, etc. All these
activities involve the muscular coordination.
2. Verbal learning:
This type of learning involves the language we speak, the
communication devices we use. Signs, pictures, symbols, words,
figures, sounds, etc, are the tools used in such activities. We use words
for communication.
3. Concept learning:
It is the form of learning which requires higher order mental processes
like thinking, reasoning, intelligence, etc. we learn different concepts
from childhood. For example, when we see a dog and attach the term
‘dog’, we learn that the word dog refers to a particular animal. Concept
learning involves two processes, viz. abstraction and generalisation.
This learning is very useful in recognising, identifying things.
4. Discrimination learning:
Learning to differentiate between stimuli and showing an appropriate
response to these stimuli is called discrimination learning. Example,
sound horns of different vehicles like bus, car, ambulance, etc.
5. Learning of principles:
Individuals learn certain principles related to science, mathematics,
grammar, etc. in order to manage their work effectively. These
principles always show the relationship between two or more
concepts. Example: formulae, laws, associations, correlations, etc.
6. Problem solving:
This is a higher order learning process. This learning requires the use
of cognitive abilities-such as thinking, reasoning, observation,
imagination, generalization, etc. This is very useful to overcome
difficult problems encountered by the people.
7. Attitude learning:
Attitude is a predisposition which determines and directs our
behaviour. We develop different attitudes from our childhood about
the people, objects and everything we know. Our behaviour may be
positive or negative depending upon our attitudes. Example: attitudes
of nurse towards her profession, patients, etc.
Theories of Learning:
Psychologists have tried to explain how people learn and why they
learn. They have conducted many experiments on animals and
children and come to certain definite conclusions which explain the
modes of learning.
The cat could see this fish. The cat was given 100 trials-ten in the
morning and ten in each afternoon for five days. The cat was fed at the
end of each experimental period and then was given nothing more to
eat until after the next session. If, succeeded in opening the door in
any trial by chance, he went to eat food (fish). A complete record was
made of the cat’s behaviour during each trial.
We all learn many skills like swimming, cycling, riding, etc., through
this method. Children learn to sit, stand, walk, and run by this method
only. However, this method involves considerable waste of time and
effort.
Learning by Conditioning:
In literal sense, conditioning means ‘getting used’ to, or ‘adjusted ‘to a
new situation, or a stimulus. It is a process of substituting the original
stimulus by a new one and connecting the response with it. There are
two types of conditioning theories:
1. Classical conditioning:
This method of conditioning got its name from the fact that, it is a
kind of learning situation that existed in the early classical
experiments of Ivan P Pavlov (1849-1936), Russian physiologist who
was awarded Nobel Prize, in 1904 for his experiments.
After the sound of the bell had been paired with food a few times, he
tested the effects of the training by measuring the amount of saliva
that flowed when he rang the bell and did not present food. He found
that some saliva was produced in response to the sound of the bell
alone. He then resumed the training-paired presentation of bell and
food a few times and then tested again with the bell alone.
As the training continued, the amount of saliva on tests with the bell
alone increased. Thus, after training the dog’s mouth watered-
salivated- whenever the bell was sounded. This is what was learned; it
is the conditioned response.
UCS<———————————à UCR
(Food) (Saliva)
↓ (Conditioning)
CS<————————————-à CR
(Bell) (Saliva)
But when the CS (bell) was paired again with UCS (food) for some
trials, the CR (salivation) recovered. This is known as ‘spontaneous
recovery’. In spontaneous recovery the dog required less number of
trials than the first time, because the association between CS and UCS
still existed in the brain of the animal.
b. Stimulus generalization:
A tendency to respond to a stimulus which is similar to original one is
called stimulus generalization, the greater the similarity, the more the
generalization. In this experiment, the dog started salivating even for
the sound of a buzzer which was similar to bell.
c. Stimulus discrimination:
When there is much difference between two stimuli, the animal can
discriminate between the two. For example, if the dog is conditioned
to salivate at the signal of red light, it will not salivate when green light
is presented.
2. Operant Conditioning:
This method of conditioning was developed by an American
psychologist BF Skinner. This theory is also known as ‘Instrumental
conditioning’, because the animals use certain operations or actions as
instruments to find solution.
Gradually, as the animal learnt the pressing of lever would give some
food, it repeated the responses very rapidly. This rapid increase in
pressing the lever is the indication of the animal conditioned to get
food.
Learning by Insight:
Many times learning proceeds by the more efficient process of trying
those methods which are seem to have a relation to solution. This is
possible by understanding or perception of the situation.
The hungry Sultan first attempted with its hands to get the banana.
Then he took one of the sticks and tried to pull the banana nearer,
then tried with other stick, but failed to reach it. By this effort, the
chimpanzee became tired and left the attempts to reach banana and
started playing with sticks.
While playing so, one of the sticks got fitted into the other and the
stick became lengthier. Immediately Sultan became elated and pulled
the banana with this long stick and ate it. This ‘sudden flash of idea’ to
reach food with longer stick was called as ‘Insight’, by Kohler.
Insight occurs, when the individual sees in a flash, the solution to his
problem or difficulty. It is not blind or stupid learning. It is an
intelligent way of learning. In many occasions people try to size up the
situation, things and arrive at a conclusion. With experience man is
able to solve problems better and sooner.
Learning by Imitation:
It is the simplest method of learning. Many of our day-to-day’s
activities are learnt by imitating others. For example, the way we eat,
drink, walk, talk, dress, etc, are all learnt by imitating others. We
observe and watch what and how other people do certain activities and
imitate them.
Psychologists like Millar and Dollard have tried to show that the
tendency to imitate is itself a learned response and if reinforced, the
individual will be more likely to continue to imitate.
Laws of Learning:
EL Thorndike has explained three laws of learning called Primary laws
and in addition to these, he has also framed 5 subsidiary laws in
connection with his trial and error learning theory.
Primary laws:
These are the most important laws, which explain the basic
aspects of learning. They are:
1. Law of readiness:
By readiness means the organism is ready to respond or act. This is
more essential prerequisite for learning.
2. Law of exercise:
This law is also known as law of frequency. Frequency refers to
number of repetitions of learning. Thorndike believed that repeated
exercising of a response strengthens its connection with stimulus.
This aspect refers to law of use and disuse, which explains that,
anything not in use will perish. So also if the response is not repeated,
its bond with stimulus gets weakened. This is also according to the
statement that ‘practice makes man perfect’.
3. Law of effect:
This law states that when a connection is accomplished by satisfying
effect- its strength is increased. By this, Thorndike meant that the
probability of its occurrence is greater. In his experiment if the hungry
cat succeeded in opening the door, would get its favourable dish to eat.
This had a positive effect on its response. Rewards always strengthen
connections between stimuli and responses, and on the other hand,
punishment weakens connections.
Secondary laws:
In addition to the three primary laws explained above, Thorndike has
given five secondary or subsidiary laws also.
This encouraged the cat to continue its effort to learn to open the door.
The secondary laws given by him support these findings. These laws
are highly relevant to the field of education. The teachers can make
use of these laws in order to make their teaching more effective.
Classical Conditioning
Classical conditioning is a learning process in which an association is made between a previously
neutral stimulus and a stimulus that naturally evokes a response.
For example, in Pavlov's classic experiment, the smell of food was the naturally occurring
stimulus that was paired with the previously neutral ringing of the bell. Once an association had
been made between the two, the sound of the bell alone could lead to a response.
For example, if you don't know how to swim and were to fall into a pool, you'd take actions to
avoid the pool.
How Classical Conditioning Works
Operant Conditioning
Operant conditioning is a learning process in which the probability of a response occurring is
increased or decreased due to reinforcement or punishment. First studied by Edward
Thorndike and later by B.F. Skinner, the underlying idea behind operant conditioning is that
the consequences of our actions shape voluntary behavior.
Skinner described how reinforcement could lead to increases in behaviors where punishment
would result in decreases. He also found that the timing of when reinforcements were delivered
influenced how quickly a behavior was learned and how strong the response would be. The
timing and rate of reinforcement are known as schedules of reinforcement.
For example, your child might learn to complete their homework because you reward them with
treats and/or praise.
How Operant Conditioning Works
Observational Learning
Observational learning is a process in which learning occurs through observing and imitating
others. Albert Bandura's social learning theory suggests that in addition to learning through
conditioning, people also learn through observing and imitating the actions of others.
Basic Principles of Social Learning Theory
As demonstrated in his classic Bobo Doll experiments, people will imitate the actions of others
without direct reinforcement. Four important elements are essential for effective observational
learning: attention, motor skills, motivation, and memory.
For example, a teen's older sibling gets a speeding ticket, with the unpleasant results of fines and
restrictions. The teen then learns not to speed when they take up driving.
The three types of learning in psychology are classical conditioning, operant conditioning, and
observational learning.
The Converger
People with this learning style have dominant abilities in the areas of Abstract Conceptualization
and Active Experimentation. They are highly skilled in the practical application of ideas. They
tend to do best in situations where there is a single best solution or answer to a problem.
The Diverger
Divergers dominant abilities lie in the areas of Concrete Experience and Reflective Observation,
essentially the opposite strengths of the Converger. People with this learning style are good at
seeing the "big picture" and organizing smaller bits of information into a meaningful whole.
Divergers tend to be emotional and creative and enjoy brainstorming to come up with new ideas.
Artists, musicians, counselors, and people with a strong interest in the fine arts, humanities, and
liberal arts tend to have this learning style.
The Assimilator
Assimilators are skilled in the areas of Abstract Conceptualization and Reflective Observation.
Understanding and creating theoretical models is one of their greatest strengths. They tend to be
more interested in abstract ideas than in people, but they are not greatly concerned with the
practical applications of theories.
Individuals who work in math and the basic sciences tend to have this type of learning style.
Assimilators also enjoy work that involves planning and research.
The Accommodator
People with this learning style are strongest in Concrete Experience and Active Experimentation.
This style is basically the opposite of the Assimilator style. Accommodators are doers; they
enjoy performing experiments and carrying out plans in the real world.
Out of all four learning styles, Accommodators tend to be the greatest risk-takers. They are good
at thinking on their feet and changing their plans spontaneously in response to new information.
When solving problems, they typically use a trial-and-error approach. People with this learning
style often work in technical fields or in action-oriented jobs such as sales and marketing.
Neutral Stimulus
A neutral stimulus is a stimulus that doesn't initially trigger a response on its own. If you hear the
sound of a fan but don't feel the breeze, for example, it wouldn't necessarily trigger a response.
That would make it a neutral stimulus.
Conditioned Stimulus
A conditioned stimulus is a stimulus that was once neutral (didn't trigger a response) but now
leads to a response. If you previously didn't pay attention to dogs, but then got bit by one, and
now you feel fear every time you see a dog, the dog has become a conditioned stimulus.
Unconditioned Response
An unconditioned response is an automatic response or a response that occurs without thought
when an unconditioned stimulus is present. If you smell your favorite food and your mouth starts
watering, the watering is an unconditioned response.
Conditioned Response
A conditioned response is a learned response or a response that is created where no response
existed before. Going back to the example of being bit by a dog, the fear you experience after the
bite is a conditioned response.
Acquisition
Acquisition is the initial stage of learning, when a response is first established and gradually
strengthened.5 During the acquisition phase of classical conditioning, a neutral stimulus is
repeatedly paired with an unconditioned stimulus.
As you may recall, an unconditioned stimulus is something that naturally and automatically
triggers a response without any learning. After an association is made, the subject will begin to
emit a behavior in response to the previously neutral stimulus, which is now known as
a conditioned stimulus. It is at this point that we can say that the response has been acquired.
Once the response has been established, you can gradually reinforce the response to make sure
the behavior is well learned.
Extinction
Extinction is when the occurrences of a conditioned response decrease or disappear. In classical
conditioning, this happens when a conditioned stimulus is no longer paired with an
unconditioned stimulus.6
For example, if the smell of food (the unconditioned stimulus) had been paired with the sound of
a whistle (the conditioned stimulus), the sound of the whistle would eventually come to evoke
the conditioned response of hunger.
However, if the smell of food were no longer paired with the whistle, eventually the conditioned
response (hunger) would disappear.
Spontaneous Recovery
Sometimes a learned response can suddenly reemerge, even after a period of extinction. This is
called spontaneous recovery.7
For example, imagine that after training a dog to salivate to the sound of a bell, you stop
reinforcing the behavior and the response becomes extinct. After a rest period during which the
conditioned stimulus is not presented, you ring the bell and the animal spontaneously recovers
the previously learned response.
If the conditioned stimulus and unconditioned stimulus are no longer associated, extinction will
return very rapidly after a spontaneous recovery.
Generalization
Stimulus generalization is the tendency for a conditioned stimulus to evoke similar responses
after the response has been conditioned.8 For example, if a dog has been conditioned to salivate
at the sound of a bell, the animal may also exhibit the same response to a sound that's similar to
the bell.
In John B. Watson's famous Little Albert Experiment, for example, a small child was
conditioned to fear a white rat. The child demonstrated stimulus generalization by also exhibiting
fear in response to other fuzzy white objects, including stuffed toys and Watson's own hair.
Discrimination
Discrimination is the ability to differentiate between a conditioned stimulus and other stimuli
that have not been paired with an unconditioned stimulus.9
For example, if a bell tone were the conditioned stimulus, discrimination would involve being
able to tell the difference between the bell tone and other similar sounds. Because the subject is
able to distinguish between these stimuli, they will only respond when the conditioned stimulus
is presented.
Fear Response
John B. Watson's experiment with Little Albert is an example of the fear response.10 The child
initially showed no fear of a white rat, but after the rat was paired repeatedly with loud, scary
sounds, the child began to cry when the rat was present.
Prior to the conditioning, the white rat was a neutral stimulus. The unconditioned stimulus was
the loud, clanging sounds, and the unconditioned response was the fear response created by the
noise.
By repeatedly pairing the rat with the unconditioned stimulus, the white rat (now the conditioned
stimulus) came to evoke the fear response (now the conditioned response).
This experiment illustrates how phobias can form through classical conditioning. In many cases,
a single pairing of a neutral stimulus (a dog, for example) and a frightening experience (being
bitten by the dog) can lead to a lasting phobia (being afraid of dogs).
Taste Aversions
Another example of classical conditioning is the development of conditioned taste aversions.
Researchers John Garcia and Bob Koelling first noticed this phenomenon when they observed
how rats that had been exposed to nausea-causing radiation developed an aversion to flavored
water after the radiation and water were presented together.11
In this example, the radiation represents the unconditioned stimulus and nausea represents the
unconditioned response. After the pairing of the two, the flavored water is the conditioned
stimulus, while nausea that formed when exposed to the water alone is the conditioned response.
Later research demonstrated that such classically conditioned aversions could be produced
through a single pairing of the conditioned stimulus and the unconditioned stimulus.
Researchers also found that such aversions can even develop if the conditioned stimulus (the
taste of the food) is presented several hours before the unconditioned stimulus (the nausea-
causing stimulus).11
Why do such associations develop so quickly? Forming such associations can have survival
benefits. If an animal eats something that makes it ill, it needs to avoid eating the same food in
the future to avoid sickness or even death.
This is an example of biological preparedness. Some associations form more readily because
they aid in survival.12
In one famous field study, researchers injected sheep carcasses with a poison that would make
coyotes sick but not kill them. The goal was to help sheep ranchers reduce the number of sheep
lost to coyote killings.
Not only did the experiment work by lowering the number of sheep killed, it also caused some of
the coyotes to develop such a strong aversion to sheep that they would actually run away at the
scent or sight of a sheep.13
Organizational Behavior
Classical conditioning can also have applications in business and marketing. For example, it can
be used to help people form favorable attitudes toward products, businesses, or brands.
While there may not be a direct link between the item and the consumer response, creating this
association may help motivate people to purchase certain products because they have developed
a favorable opinion of them due to classical conditioning.
Criticisms of Classical
Conditioning
Some psychologists maintain that classical conditioning represents a reductive, mechanical
explanation for some behaviors. Some other criticisms of classical conditioning center on the fact
that:
Classical conditioning does not take human individuality and free will into account
It generally does not predict human behavior; people can form associations but still not
act upon them
Many different factors can impact the associations and outcomes
People can choose to not act on the associations they have made through classical
conditioning
However, the approach still holds great fascination for researchers and relevance in modern
psychology.
In reality, people do not respond exactly like Pavlov's dogs. There are, however, numerous real-
world applications for classical conditioning. For example, many dog trainers use classical
conditioning techniques to help people train their pets.
These techniques are also useful for helping people cope with phobias or anxiety problems.
Therapists might, for example, repeatedly pair something that provokes anxiety with relaxation
techniques in order to create an association.
Teachers can apply classical conditioning in the class by creating a positive classroom
environment to help students overcome anxiety or fear. Pairing an anxiety-provoking situation,
such as performing in front of a group, with pleasant surroundings helps the student learn new
associations. Instead of
Types of Behaviors
Skinner distinguished between two different types of behaviors
Respondent behaviors are those that occur automatically and reflexively, such as pulling
your hand back from a hot stove or jerking your leg when the doctor taps on your knee.
You don't have to learn these behaviors. They simply occur automatically and
involuntarily.
Operant behaviors, on the other hand, are those under our conscious control. Some may
occur spontaneously and others purposely, but it is the consequences of these actions that
then influence whether or not they occur again in the future. Our actions on the
environment and the consequences of that action make up an important part of
the learning process.
While classical conditioning could account for respondent behaviors, Skinner realized that it
could not account for a great deal of learning. Instead, Skinner suggested that operant
conditioning held far greater importance.
Skinner invented different devices during his boyhood and he put these skills to work during his
studies on operant conditioning. He created a device known as an operant conditioning chamber,
often referred to today as a Skinner box. The chamber could hold a small animal, such as a rat or
pigeon. The box also contained a bar or key that the animal could press in order to receive a
reward.
In order to track responses, Skinner also developed a device known as a cumulative recorder.
The device recorded responses as an upward movement of a line so that response rates could be
read by looking at the slope of the line.
Components of Operant
Conditioning
There are several key concepts in operant conditioning. The type of reinforcement or punishment
that is used can have an effect on how the individual responds and the effect of conditioning.
There are four types of operant conditioning that can be utilized to change behavior: positive
reinforcement, negative reinforcement, positive punishment, and negative punishment.
Reinforcement in Operant Conditioning
Reinforcement is any event that strengthens or increases the behavior it follows. There are two
kinds of reinforcers. In both of these cases of reinforcement, the behavior increases.
1. Positive reinforcers are favorable events or outcomes that are presented after the
behavior. In positive reinforcement situations, a response or behavior is strengthened by
the addition of praise or a direct reward. If you do a good job at work and your manager
gives you a bonus, that bonus is a positive reinforcer.
2. Negative reinforcers involve the removal of an unfavorable events or outcomes after the
display of a behavior. In these situations, a response is strengthened by the removal of
something considered unpleasant. For example, if your child starts to scream in the
middle of a restaurant, but stops once you hand them a treat, your action led to the
removal of the unpleasant condition, negatively reinforcing your behavior (not your
child's).
Reinforcement in Conditioning Behavior
Observational Learning In
Psychology: Example
Behaviors
Observational learning involves acquiring skills or new or changed behaviors
through watching the behavior of others.
The person or actor performing the action that the observational learner
replicates is called a model.
Background
Observational learning, otherwise known as vicarious learning, is the
acquisition of information, skills, or behavior through watching others
perform, either directly or through another medium, such as video.
His theory, social learning theory, stresses the importance of observation and
modeling of behaviors, attitudes, and the emotional reactions of others.
Even infants may start imitating the mouth movements and facial expressions
of the adults around them.
These variables can include how similar the model is to the observer and
the observer’s current mood.Humans, Bandura (1985) proposed, are
likely to pay attention to the behaviors of models that are high-status,
talented, intelligent, or similar to the learner in some way.
Forexample, someone seeking to climb the corporate ladder may
observe the behavior of their managers and the vice presidents of their
company, and try to mimic their behavior (Debell, 2021).
Retention
Attentionin itself, however, is not enough to learn a new behavior.
Observers Must also retain, or remember, the behavior at a later time.
In order to increase the chances of retention, the observer can
structure the information in a way that is easy to remember.
Motor Reproduction
After retention comes the ability to actually perform a behavior in real-
life. Often, producing a new behavior can require hours of practice in
order to obtain the necessary skills to do so.
Thus, the process of reproduction is one that can potentially take years
to craft and perfect (Debell, 2021).
Motivation
Finally,
all learning requires, to some extent, personal motivation. Thus,
in observational learning, an observer must be motivated to produce
the desired behavior.
After watching this behavior, the researchers gave the children a bobo doll
identical to the one in the video.
The researchers found that children were more likely to mimic violent
behaviors when they observed the model receiving a reward, or when no
consequences occurred.
Alternatively, children who observed the model being punished for their
violence showed less violence toward the doll (Debell, 2021).
A school child may learn to write cursive letters through observing their
teacher write them on the board.
Children may learn to play hide and seek by seeing other children
playing the game and being rewarded in the form of entertainment
Children may also learn to say swear words after watching other
children say swear words and gain social status.
A child may be able to put on roller skates and stand on them without
explicit instruction.
A child may avoid stepping on ice after seeing another child fall in front
of them.
Meanwhile, parents who want their children to eat healthily can in themselves
eat healthily and exercise, as well as spend time engaging in physical fitness
activities together.
Observational learning argues, in all, that children tend to copy what parents
do above what they say (Daffin, 2021).
Observational learning has also been used to explain how antisocial behaviors
develop. For example, research suggests that observational learning is a
reason why many abused children grow up to become abusers themselves
(Murrel, Christoff, & Henning, 2007).
Abused children tend to grow up witnessing their parents deal with anger and
frustration through violent and aggressive acts, often learning to behave in
that manner themselves.
Some studies have also suggested that violent television shows may also have
antisocial effects, though this is a controversial claim (Kirsh, 2011).
Observational Learning and Behavioral Modification
Observational learning can be used to change already learned behaviors, for
both positive and negative.
Bandura asserted that, if all behaviors are learned by observing others and
people can model their behavior on that of those around them, then
undesirable behaviors can be altered or relearned in the same way.
By seeing the model interact nicely with the fear-evoking stimulus, the fear
should subside. This method of behavioral modification is widely used in
clinical, business, and classroom situations (Daffin, 2021).
If the student can solve the problem, no further action is needed; however, if
the student struggles, a teacher may use one of four types of prompts — verbal,
gestural, modeling, or physical — to assist the student. Similarly, a trainer may
show a trainee how to use a computer program to run a register.
As with before, the trainer can use prompt delays and prompts to test the level
of learning the employee has gained.
Reinforcers can then be delivered through social support after the trainee has
successfully completed the task themself (Daffin, 2021).
Pavlov conditioned a number of dogs by pairing food with the tone of a bell.
After several repetitions, he was able to trigger his dogs to salivate by ringing
the bell, even in the absence of food.
While classical and operant conditioning may rely on trial and error alone as a
means of changing behavior, observational conditioning creates room for
observing a model, whose actions someone can replicate.
This can result in a more controlled and ultimately more efficient learning
process for all involved (Daffin, 2021).
Pavlov predicted the dogs would salivate in response to the food in front of
them, but he noticed that his dogs would begin to salivate whenever they
heard the footsteps of his assistant, who was bringing them the food.
When Pavlov discovered that any object or event that the dogs learned to
associate with food (such as the lab assistant) would trigger the same
response, he realized that he had made an important scientific discovery.
Accordingly, he devoted the rest of his career to studying this type of learning.
Table of Contents
Pavlov developed some rather unfriendly technical terms to describe this process:
NeutralStimulus (NS): A stimulus that initially does not elicit a
particular response or reflex action. In other words, before any
conditioning takes place, the neutral stimulus has no effect on the
behavior or physiological response of interest. For example, in
Pavlov’s experiment, the sound of a metronome was a neutral
stimulus initially, as it did not cause the dogs to salivate.
Unconditioned Stimulus (UCS): This is a stimulus that naturally and
automatically triggers a response without any learning needed. In
Pavlov’s experiment, the food was the unconditioned stimulus as it
automatically induced salivation in the dogs.
Pavlov showed that dogs could be conditioned to salivate at the sound of a bell
if that sound was repeatedly presented at the same time that they were given
food.
Pavlov and his studies of classical conditioning have become famous since his
early work between 1890-1930. Classical conditioning is “classical” in that it is
the first systematic study of the basic laws of learning (also known as
conditioning).
Pavlov’s dogs were individually situated in secluded environments, secured
within harnesses. A food bowl was positioned before them, and a device was
employed to gauge the frequency of their salivary gland secretions.
First, the dogs were presented with the food, and they salivated. The food was
the unconditioned stimulus and salivation was an unconditioned (innate)
response. (i.e., a stimulus-response connection that required no learning).
He called this the law of temporal contiguity. If the time between the
conditioned stimulus (bell) and the unconditioned stimulus (food) is too great,
then learning will not occur.
Over time, the dog stops associating the sound of the bell with the food, and
the conditioned response (salivation) weakens and eventually disappears.
Spontaneous recovery
Pavlov noted the occurrence of “spontaneous recovery,” where the conditioned
response can briefly reappear when the conditioned stimulus is presented
after a rest period, even though the response has been extinguished.
Generalization
The principle of generalization suggests that after a subject has been
conditioned to respond in a certain way to a specific stimulus, the subject will
also respond in a similar manner to stimuli that are similar to the original one.
However, the response tends to be more pronounced when the new stimulus
closely resembles the original one used in conditioning.
This relationship between the similarity of the stimulus and the strength of the
response is known as the generalization gradient.
For instance, a brand may pair its product with appealing stimuli (like
enjoyable music or attractive visuals) to create a positive emotional
response in consumers, which then gets associated with the product.
By the 1920s, John B. Watson had left academic psychology, and other behaviorists were
becoming influential, proposing new forms of learning other than classical conditioning. Perhaps
the most important of these was Burrhus Frederic Skinner. Although, for obvious reasons, he is
more commonly known as B.F. Skinner.
Skinner’s views were slightly less extreme than those of Watson (1913). Skinner believed that
we do have such a thing as a mind, but that it is simply more productive to study observable
behavior rather than internal mental events.
The work of Skinner was rooted in the view that classical conditioning was far too simplistic to
be a complete explanation of complex human behavior. He believed that the best way to
understand behavior is to look at the causes of an action and its consequences. He called this
approach operant conditioning.
BF Skinner: Operant Conditioning
Skinner is regarded as the father of Operant Conditioning, but his work was based
on Thorndike’s (1898) law of effect. According to this principle, behavior that is followed by
pleasant consequences is likely to be repeated, and behavior followed by unpleasant
consequences is less likely to be repeated.
Skinner introduced a new term into the Law of Effect – Reinforcement. behavior that is
reinforced tends to be repeated (i.e., strengthened); behavior that is not reinforced tends to die
out or be extinguished (i.e., weakened).
Skinner (1948) studied operant conditioning by conducting experiments using animals which he
placed in a “Skinner Box ” which was similar to Thorndike’s puzzle box.
A Skinner box, also known as an operant conditioning chamber, is a device used to
objectively record an animal’s behavior in a compressed time frame. An animal can be
rewarded or punished for engaging in certain behaviors, such as lever pressing (for rats)
or key pecking (for pigeons).
Skinner identified three types of responses, or operant, that can follow behavior.
Neutral operants: responses from the environment that neither increase nor decrease the
probability of a behavior being repeated.
Reinforcers: Responses from the environment that increase the probability of a behavior
being repeated. Reinforcers can be either positive or negative.
Punishers: Responses from the environment that decrease the likelihood of a behavior
being repeated. Punishment weakens behavior.
We can all think of examples of how our own behavior has been affected by reinforcers and
punishers. As a child, you probably tried out a number of behaviors and learned from their
consequences.
For example, when you were younger, if you tried smoking at school, and the chief consequence
was that you got in with the crowd you always wanted to hang out with, you would have been
positively reinforced (i.e., rewarded) and would be likely to repeat the behavior.
If, however, the main consequence was that you were caught, caned, suspended from school and
your parents became involved you would most certainly have been punished, and you would
consequently be much less likely to smoke now.
Positive Reinforcement
Positive reinforcement is a term described by B. F. Skinner in his theory of operant conditioning.
In positive reinforcement, a response or behavior is strengthened by rewards, leading to the
repetition of desired behavior. The reward is a reinforcing stimulus.
Primary reinforcers are stimuli that are naturally reinforcing because they are not learned and
directly satisfy a need, such as food or water.
Secondary reinforcers are stimuli that are reinforced through their association with a primary
reinforcer, such as money, school grades. They do not directly satisfy an innate need but may be
the means. So a secondary reinforcer can be just as powerful a motivator as a primary reinforcer.
Skinner showed how positive reinforcement worked by placing a hungry rat in his Skinner box.
The box contained a lever on the side, and as the rat moved about the box, it would accidentally
knock the lever. Immediately it did so that a food pellet would drop into a container next to the
lever.
The rats quickly learned to go straight to the lever after a few times being put in the box. The
consequence of receiving food, if they pressed the lever, ensured that they would repeat the
action again and again.
This method incentivizes the less desirable behavior by associating it with a desirable outcome,
thus strengthening the less favored behavior.
Negative Reinforcement
Negative reinforcement is the termination of an unpleasant state following a response.
This is known as negative reinforcement because it is the removal of an adverse stimulus which
is ‘rewarding’ to the animal or person. Negative reinforcement strengthens behavior because it
stops or removes an unpleasant experience.
For example, if you do not complete your homework, you give your teacher £5. You will
complete your homework to avoid paying £5, thus strengthening the behavior of
completing your homework.
Skinner showed how negative reinforcement worked by placing a rat in his Skinner box and then
subjecting it to an unpleasant electric current which caused it some discomfort. As the rat moved
about the box it would accidentally knock the lever.
Immediately it did so the electric current would be switched off. The rats quickly learned to go
straight to the lever after a few times of being put in the box. The consequence of escaping the
electric current ensured that they would repeat the action again and again.
In fact, Skinner even taught the rats to avoid the electric current by turning on a light just before
the electric current came on. The rats soon learned to press the lever when the light came on
because they knew that this would stop the electric current from being switched on.
These two learned responses are known as Escape Learning and Avoidance Learning.
Like reinforcement, punishment can work either by directly applying an unpleasant stimulus like
a shock after a response or by removing a potentially rewarding stimulus, for instance, deducting
someone’s pocket money to punish undesirable behavior.
Note: It is not always easy to distinguish between punishment and negative reinforcement.
Creates fear that can generalize to undesirable behaviors, e.g., fear of school.
Doesnot necessarily guide you toward desired behavior – reinforcement tells you what to
do, and punishment only tells you what not to do.
2. Negative Reinforcement: If you notice your team working together effectively and
exhibiting excellent team spirit during a tough training session, you might end the
training session earlier than planned, which the team perceives as a relief. They
understand that teamwork leads to positive outcomes, reinforcing team behavior.
4. Positive Reinforcement: Training a cat to use a litter box can be achieved by giving it
a treat each time it uses it correctly. The cat will associate the behavior with the reward
and will likely repeat it.
5. Negative Punishment: If teenagers stay out past their curfew, their parents might take
away their gaming console for a week. This makes the teenager more likely to respect
their curfew in the future to avoid losing something they value.
6. Ineffective Punishment: Your child refuses to finish their vegetables at dinner. You
punish them by not allowing dessert, but the child still refuses to eat vegetables next
time. The punishment seems ineffective.
7. Premack Principle Application: You could motivate your child to eat vegetables by
offering an activity they love after they finish their meal. For instance, for every
vegetable eaten, they get an extra five minutes of video game time. They value video
game time, which might encourage them to eat vegetables.
For every 10 minutes a person spends on household chores, they can spend
5 minutes on a favorite hobby.
A child can choose between taking out the trash or washing the dishes.
Giving them the choice makes them more likely to complete the chore
willingly.
One of the most famous of these experiments is often colloquially referred to as “Superstition in
the Pigeon.”
The Experiment:
1. Pigeons were brought to a state of hunger, reduced to 75% of their well-fed weight.
2. They were placed in a cage with a food hopper that could be presented for five seconds
at a time.
3. Instead of the food being given as a result of any specific action by the pigeon, it was
presented at regular intervals, regardless of the pigeon’s behavior.
Observation:
1. Over time, Skinner observed that the pigeons began to associate whatever random
action they were doing when food was delivered with the delivery of the food itself.
2. This led the pigeons to repeat these actions, believing (in anthropomorphic terms) that
their behavior was causing the food to appear.
Findings:
1. In most cases, pigeons developed different “superstitious” behaviors or rituals. For
instance, one pigeon would turn counter-clockwise between food presentations, while
another would thrust its head into a cage corner.
2. These behaviors did not appear until the food hopper was introduced and presented
periodically.
3. These behaviors were not initially related to the food delivery but became linked in the
pigeon’s mind due to the coincidental timing of the food dispensing.
4. The behaviors seemed to be associated with the environment, suggesting the pigeons
were responding to certain aspects of their surroundings.
5. The rate of reinforcement (how often the food was presented) played a significant role.
Shorter intervals between food presentations led to more rapid and defined
conditioning.
Superstitious Behavior:
The pigeons began to act as if their behaviors had a direct effect on the presentation of food,
even though there was no such connection. This is likened to human superstitions, where rituals
are believed to change outcomes, even if they have no real effect.
For example, a card player might have rituals to change their luck, or a bowler might make
gestures believing they can influence a ball already in motion.
Conclusion:
This experiment demonstrates that behaviors can be conditioned even without a direct cause-and-
effect relationship. Just like humans, pigeons can develop “superstitious” behaviors based on
coincidental occurrences.
This study not only sheds light on the intricacies of operant conditioning but also draws parallels
between animal and human behaviors in the face of random reinforcements.
Schedules Of Reinforcement
Imagine a rat in a “Skinner box.” In operant conditioning, if no food pellet is delivered
immediately after the lever is pressed then after several attempts the rat stops pressing the lever
(how long would someone continue to go to work if their employer stopped paying them?). The
behavior has been extinguished.
Behaviorists discovered that different patterns (or schedules) of reinforcement had different
effects on the speed of learning and extinction. Ferster and Skinner (1957) devised different
ways of delivering reinforcement and found that this had effects on
1. The Response Rate – The rate at which the rat pressed the lever (i.e., how hard the rat
worked).
2. The Extinction Rate – The rate at which lever pressing dies out (i.e., how soon the rat gave
up).
Skinner found that the type of reinforcement which produces the slowest rate of extinction (i.e.,
people will go on repeating the behavior for the longest time without reinforcement) is variable-
ratio reinforcement. The type of reinforcement which has the quickest rate of extinction is
continuous reinforcement.
Response rate is FAST
Extinction rate is SLOW (very hard to extinguish because of
unpredictability)
This is not as simple as it sounds — always reinforcing desired behavior, for example, is
basically bribery.
There are different types of positive reinforcements. Primary reinforcement is when a reward
strengths a behavior by itself. Secondary reinforcement is when something strengthens a
behavior because it leads to a primary reinforcer.
Examples of behavior modification therapy include token economy and behavior shaping.
Token Economy
Token economy is a system in which targeted behaviors are reinforced with tokens (secondary
reinforcers) and later exchanged for rewards (primary reinforcers).
Tokens can be in the form of fake money, buttons, poker chips, stickers, etc. While the rewards
can range anywhere from snacks to privileges or activities. For example, teachers use token
economy at primary school by giving young children stickers to reward good behavior.
Token economy has been found to be very effective in managing psychiatric patients. However,
the patients can become over-reliant on the tokens, making it difficult for them to adjust to
society once they leave prison, hospital, etc.
Staff implementing a token economy program have a lot of power. It is important that staff do
not favor or ignore certain individuals if the program is to work. Therefore, staff need to be
trained to give tokens fairly and consistently even when there are shift changes such as in prisons
or in a psychiatric hospital.
Behavior Shaping
A further important contribution made by Skinner (1951) is the notion of behavior shaping
through successive approximation. Skinner argues that the principles of operant conditioning can
be used to produce extremely complex behavior if rewards and punishments are delivered in
such a way as to encourage move an organism closer and closer to the desired behavior each
time.
In shaping, the form of an existing response is gradually changed across successive trials towards
a desired target behavior by rewarding exact segments of behavior.
To do this, the conditions (or contingencies) required to receive the reward should shift each time
the organism moves a step closer to the desired behavior.
According to Skinner, most animal and human behavior (including language) can be explained
as a product of this type of successive approximation.
Educational Applications
In the conventional learning situation, operant conditioning applies largely to issues of class and
student management, rather than to learning content. It is very relevant to shaping skill
performance.
For example, if a teacher wanted to encourage students to answer questions in class they should
praise them for every attempt (regardless of whether their answer is correct). Gradually the
teacher will only praise the students when their answer is correct, and over time only exceptional
answers will be praised.
Unwanted behaviors, such as tardiness and dominating class discussion can be extinguished
through being ignored by the teacher (rather than being reinforced by having attention drawn to
them). This is not an easy task, as the teacher may appear insincere if he/she thinks too much
about the way to behave.
Learning Process
Classical conditioning involves learning through associating stimuli resulting in involuntary
responses, while operant conditioning focuses on learning through consequences, shaping
voluntary behaviors.
Over time, the person responds to the neutral stimulus as if it were the unconditioned
stimulus, even when presented alone. The response is involuntary and automatic.
An example is a dog salivating (response) at the sound of a bell (neutral stimulus) after
it has been repeatedly paired with food (unconditioned stimulus).
For instance, if a child gets praised (pleasant consequence) for cleaning their room
(behavior), they’re more likely to clean their room in the future.
Conversely, if they get scolded (unpleasant consequence) for not doing their
homework, they’re more likely to complete it next time to avoid the scolding.
Timing of Stimulus & Response
The timing of the response relative to the stimulus differs between classical and operant
conditioning:
For example, in Pavlov’s classic experiment, the dogs started to salivate (response)
after they heard the bell (stimulus) because they associated it with food.
Operant Conditioning (response before the stimulus): In this form of conditioning, the
response generally occurs before the consequence (which acts as the stimulus for
future behavior).
The anticipated consequence influences the behavior or what follows it. It is a more
active form of learning, where behaviors are reinforced or punished, thus influencing
their likelihood of repetition.
Summary
Looking at Skinner’s classic studies on pigeons’ / rat’s behavior we can identify some of the
major assumptions of the behaviorist approach.
• The major influence on human behavior is learning from our environment. In the
Skinner study, because food followed a particular behavior the rats learned to repeat that
behavior, e.g., operant conditioning.
• There is little difference between the learning that takes place in humans and that in
other animals. Therefore research (e.g., operant conditioning) can be carried out on
animals (Rats / Pigeons) as well as on humans. Skinner proposed that the way humans
learn behavior is much the same as the way the rats learned to press a lever.
So, if your layperson’s idea of psychology has always been of people in laboratories wearing
white coats and watching hapless rats try to negotiate mazes in order to get to their dinner, then
you are probably thinking of behavioral psychology.
Behaviorism and its offshoots tend to be among the most scientific of the psychological
perspectives. The emphasis of behavioral psychology is on how we learn to behave in certain
ways.
We are all constantly learning new behaviors and how to modify our existing behavior.
behavioral psychology is the psychological approach that focuses on how this learning takes
place.
Critical Evaluation
Operant conditioning can be used to explain a wide variety of behaviors, from the process of
learning, to addiction and language acquisition. It also has practical applications (such as token
economy) which can be applied in classrooms, prisons and psychiatric hospitals.
However, operant conditioning fails to take into account the role of inherited and cognitive
factors in learning, and thus is an incomplete explanation of the learning process in humans and
animals.
For example, Kohler (1924) found that primates often seem to solve problems in a flash of
insight rather than be trial and error learning. Also, social learning theory (Bandura, 1977)
suggests that humans can learn automatically through observation rather than through personal
experience.
The use of animal research in operant conditioning studies also raises the issue of extrapolation.
Some psychologists argue we cannot generalize from studies on animals to humans as their
anatomy and physiology is different from humans, and they cannot think about their experiences
and invoke reason, patience, memory or self-comfort.