Operant Conditioning
Download
Report
Transcript Operant Conditioning
Operant Conditioning
Complex Learning
Why do we learn new behaviors?
Classical conditioning only deals with reflex
responses that we already possess.
Most of our behaviors are voluntary.
Volitional. Stimulated by something in our
environment.
Operant Conditioning
Defined as - the form of learning concerned
with changes in emitted responses as a
function of their consequences.
Origins of Operant
Conditioning
Edward Thorndike
Instrumental
Conditioning
“Law of Effect”
Satisfying outcome
Unsatisfactory outcome
Outcomes of Thorndike’s Work
How Long - the length
it declined
ofAs
time
it tooklearning
the cat to
was taking
escape
fromplace.
the puzzle
This change in
box.
performance
represented a change
in behavior from
experience.
Learning
Question
In Thorndike’s terms, what sort of things
give you satisfaction? What things produce
dissatisfaction? Why?
Edward Thorndike
His research provided a foundation for the
study of “non-reflexive” learning.
He drew a connection between action and its
outcomes.
B. F. Skinner
Skinner coined the
term “operant”.
Disagreed with the
“soft” concepts of
Thorndike’s
“satisfying” and
“unsatisfactory”
outcome(s)
B. F. Skinner
Operant Conditioning
replaced Thorndike’s term
“instrumental learning”
Emitted behavior is now
called “operant responses”
Classical conditioning is
now called ‘respondent
conditioning.
The Skinner Box or “autoenvironmental chamber”
Skinner Box in Action
Zack Florin '99 using a Skinner box to shape a rat's behavior
Reinforcment
Primary reinforcers - food, water, shelter.
Those innate biological needs.
Secondary reinforcers (Conditioned
reinforcers) - something that will provide a
primary reinforcer. (money, poker chips etc.)
Primary vs. Secondary
Which of the following are secondary
reinforcers:
quarters spilling from a slot machine,
a winner’s blue ribbon,
a piece of candy,
an A on an exam,
frequent-flyer miles.
Reinforcement
Negative Reinforcer - an aversive stimulus
which serves to decrease the probability of
the response in the future.
Positive Reinforcer - a stimulus which when
applied increases the probability of the
response in the future.
Contingencies of
Reinforcement
According to Skinner the relationship
between a response and a reinforcer is a
contingency.
One type of contingency is “reinforcement”
Desired change in behavior
Type of reinforcer
Increases response
Decrease response
Positive reinforcer
POSITIVE
REINFORCEMENT
OMISSION
NEGATIVE
REINFORCEMENT
PUNISHMENT
Negative reinforcer
(escape, avoidance)
(withholding positive
reinforcer)
Shaping
Some learning does not
occur in a single event.
A series of successive
steps leads to a learned
behavior.
Playing the piano,
swimming etc.
Applying the Principles
When asked choose the best alternative and
explain why.
You want your 2-year-old to ask for water with a
word instead of a grunt. Should you give him
water when he says “wa-wa” or wait until his
pronunciation improves.
Applying the Principles
When asked choose the best alternative and
explain why.
Your roommate keeps interrupting your studying
even though you have asked her/him to stop.
Should you ignore her/him completely or
occasionally respond for the sake of good
manners?.
Applying the Principles
When asked choose the best alternative and
explain why.
Your father, who rarely writes to you, has finally
sent a letter. Should you reply quickly or wait a
while so he will know how it feels to be
ignored?.
Extinction
What happens when
the reinforcement
stops.
Extinction - in operant
conditioning, a drop I
responding when
reinforcement is
discontinued.
Schedules of Reinforcement
Continuous reinforcement - every response is
followed by a reinforcer. (FR1 schedule)
Partial reinforcement - a contingency of
reinforcement in which every response does
not get a reinforcer.
Fixed Interval Schedule
Referred to as FI x reinforcement
contingency defined by
the amount of time that
must pass since the
previous reinforcer.
Based on time.
Example: pay checks
Fixed Ratio Schedule
Referred to as FR x reinforcement
contingency defined by
the number of
responses the organism
must make in order to
get a reinforcer.
Example: piece work.
Variable Interval Schedule
Referred to as VI x - a
reinforcement
contingency defined by
the average time
interval which must
elapse since the last
reinforcer.
Example: Quality
Control
Variable Ratio Schedule
Referred to as VR x - a
reinforcement
contingency defined in
terms of the average
number of responses
required to receive a
reinforcer.
Example: Slot Machine
Non-Contingent Reinforcement
Random
“reinforcement”
Development of what
Skinner called
‘superstition’ in the
pigeon.
Applying Conditioning
We must always keep in mind that all this is
done to match the goals of psychology.
Behavior Modification.
Mary Cover Jones - the mother of behavior
therapy
Aversive
Positive
Controls
Punishment
Most used and most
misunderstood
Occurs after the
‘offense’ has taken
place.
Requires “contiguity”
Encourages avoidance
behaviors.
Negative Reinforcement
Autonomic Conditioning
Neal Miller and Leo
DiCara
‘proprioceptive
feedback
Biological Constraints
Some unanswered
questions:
Equipotentiality
premise
Ethology
Species-specific
behavior
Critical period
Preparedness
• The
The
premise
that
principles
Behaviors
which
are
A
study
period
of
during
the
behavior
of
A
concept
developed
by
ofdevelopment
conditioning
willmembers
apply
characteristic
of
all
animals
in
their
natural
where
there
Martin
Seligman
to
toare
any
response
and any
of
a
particular
species.
environment.
optimal
periods
for
describe
how
species.
(instincts)
learning.
physiological
structure
influences the
occurrence of behavior
Biological Constraints
Degree of biological preparedness
Prepared
Species-specific
behavior
Unprepared
Bait Shyness
Classical
and operant
conditioning
Contraprepared
Unlearnable
Associations