Chapter 8, Learning



Unit 6, LEARNING notes, pp. 215-253

Updated for 2011; goes with Ch. 8 in 2007 book)

Adaptability: capacity to learn new behaviors that help us cope with changing circumstances

Learning: a relatively permanent change in behavior based on experience

Habituation: an organism’s decreasing response to a stimulus with repeated exposure to it.

Associative learning: forming an association between cause and effect

Example: associating the sound of thunder with the sight of lightning

Conditioning: the process of learning associations

Operant conditioning: learning to associate a response with its consequence

Observational learning: learning from the examples of others

Classical conditioning: learning to associate to stimuli to anticipate events

Behaviorism: Watson’s view that psychology should be an objective science based on observable behavior

Classical conditioning

What is classical conditioning and how did Pavlov’s work influence behaviorism?

IVAN PAVLOV, JOHN WATSON page 218

Developed by Ivan Pavlov using dog salivation experiments

Stimuli is associated with an involuntary response

Respondent behavior: an automatic response to certain stimuli

Unconditioned response (UCR or UR) the normal response; in Pavlov’s experiment it was the salivation of the dog.

Unconditioned stimulus (UCS or US) the stimulus that triggers a normal response

Conditioned response (CR) the response that is learned; in Pavlov’s exp. it was salivation to a metronome or turning fork and other sounds.

Note: the UR and CR are almost always the same.

Conditioned stimulus (CS) a neutral stimulus that triggers a learned response; in Pavlov’s experiment it was the sound

The US should precede the CS by about ½ second.

Acquisition, 220

What are the five major processes in CC?

There are five major processes with classical conditioning

1. Acquisition: the initial formation of the association between the CS and the CR

Higher order conditioning: the CS in one experience is paired with a new NS, creating a second CS. Ex: an animal that has learned a tone produces food may learn that a light signals the presentation of the tone.

2. Extinction: if the UCS is not presented, after a few times the organism loses receptivity to the CS

3. Spontaneous recovery: however, if the UCS is presented again with the CS, the organism begins to respond again

4. Generalization: the tendency to respond similarly to similar stimuli;

5. Discrimination: the ability to distinguish between different stimuli

___________________________________________________________________

Updating Pavlov, 223

Do cognitive processes and biological constraints affect classical conditioning?

Behaviorist learning doesn’t allow us to generalize from one response to another, nor from one species to another

Cognitive processes

1. Rescorla and Wagner (1972): animals learn to “expect” an unconditioned stimulus; this shows cognition at work: the animal learns the predictability of a second associated event after the first

2. Conditioning an alcoholic with a nauseating drink might not work because they are “aware” of what causes the nausea---the drink, not alcohol

Martin Seligman found that dogs given repeated shocks with no opportunity to avoid them developed a passive resignation called learned helplessness.

Some dogs just never gave up. He developed the idea of learned optimism and now teaches people to be more optimistic and less helpless.

Biological predispositions

Garcia and Koelling:

1. Animals can learn to avoid a drink that will make them sick, but not when it’s announced by a noise; so Pavlov was wrong in claiming that any stimulus could serve as a conditioned stimulus.

2. We are biologically prepared to learn certain associations and not others: we learn to fear snakes, but not flowers

3. Taste aversions result from biology. The smell and taste of a nauseating food become the CS for sickness.

4. Secondary disgust: an aversion that reminds a person of something that is considered disgusting in its own right. Ex. You won’t eat a piece of chocolate formed like dog feces.

The biopsychosocial influences on learning (see fig. 6.9, p. 225): genetic predispositions, previous experiences, predictability of associations, culturally learned preferences, and motivations all influence learning.

Many phobias are learned through classical conditioning. Systematic desensitization can also unlearn these fears. (see pp. 611-612)

Pavlov’s legacy, 226

Virtually all organisms learn by classical conditioning.

Pavlov showed how the process can be studied objectively.

_____________________________________________________________________

Practical applications

What have been some applications of classical conditioning?

1. Treating drug addiction

2. Treating emotional illness

5. Increasing the immune system’s responsiveness (Robert Ader’s rat experiment: rats who liked saccharine inadvertently were conditioned to lower their immune system when the sweet substance was paired with a chemical that made them nauseous)

6. Pavlov’s worked paved the way for John Watson’s Little Albert Study:

He conditioned Albert to fear (CR) a white rat (NS) when the rat’s presence was paired with the sound of a loud noise (UCS). It only took seven pairings.

After five days Albert became afraid of other fury white animals (generalization).

Watson showed how emotions could be conditioned. An opposite effect to the Albert experiment can be created: you can learn not to fear things.

Watson later worked in advertising and invented the concept of the “coffee break”.

II. Operant conditioning BF SKINNER page 228

What is operant conditioning, and how does it differ from classical conditioning?

Developed by B. F. Skinner where subjects voluntarily respond a certain way depending on the consequences (reward or punishment)

Respondent behavior: it occurs as an automatic response to some stimulus

Operant behavior: the behavior that acts upon a situation and produces consequences

Ex: eating in your parent’s living makes them mad; your eating behavior may change depending on the kind of consequences you want (yelling or not) the situation to have

Skinner’s experiments, 229

Law of effect: EL Thordinke’s idea that behavior that is rewarded is more likely to recur (one of Psychology’s few laws)

Operant chamber: Skinner box used to reward animal behavior; this behavioral technology revealed the principles of behavior control

Shaping: reinforcers guide behavior; gradually rewarding behavior as it approaches the desired target; important because organisms rarely perform the desired behavior at first.

Successive approximations: rewarding responses that are closer to the target behavior

Discriminative stimulus: it elicits a response after association with reinforcement; these are used to help animals form concepts

Types of Reinforcers, 230

What are the basic types of reinforcers?

Reinforcers: anything that increases the chance of a behavior occurring again

1. Positive: things the organism desires

Ex: rewards like appraisal, money, and food

2. Negative: the removal (-) of aversive events

Freeing someone from prison

Taking a medicine to relieve a cold

3. Primary things that satisfy inborn biological needs

Ex: food, water, warmth

4. Secondary (also called “conditioned”) learned things that are strengthened by association with primary reinforcers

Ex: money that can be used to buy the primary reinforcer, food

Ex: praise, high grades, smiles which can lead to the primary emotion, happiness

Immediate and delayed page 231

Unlike rats, humans do respond to delayed reinforcement

Four year olds have been shown to delay reinforcement; example: Walter Mischel’s marshmallow experiment)

But drug users succumb to immediate gratification; the later ill effects often not enough to prevent addiction

Teens fall to immediate gratification of sex over safe or saved sex

Schedules of reinforcement pages 232-233

How do different reinforcement schedules affect behavior?

a. Continuous

A quick method of learning in which the desired behavior is reinforced every time

when reinforcement stops extinction can quickly happen; a teacher awarding points for every assignment is an example.

b. Partial reinforcing a behavior parts of the time; the acquisition is slower but more resistant to change

c. Fixed ratio reinforcement after a fixed number of response

Ex: getting candy after washing the floor every 3 times

d. Variable ratio

Ex: getting candy after the third floor washing, then the 2nd, then the 5th

Ratio schedules produce high rates of responding; reinforcers increase as the numbers of responses increase.

e. Fixed interval reinforcement after a fixed amount of time

Ex: getting candy three hours after every time the floor is washed

f. Variable interval reinforcement after an unpredictable amount of time

Ex: getting candy two hours after the floor is washed, then 3, then 5

Variable schedules produce more consistent responding.

Punishment, 234

How does punishment affect behavior?

The opposite of reinforcement: the goal is to decrease behavior, not to increase it

John Darley: swift and sure is the most effective punishment

Positive punishment: administer something the subject doesn’t like

Example: spanking; parking ticket

Negative punishment withdrawal of something desirable

Example: timeout (the child can get out of the time out if they agree to comply with parent requests); revoking privileges

Punishment pros and cons

1. successfully stops behavior

2. drawback: the behavior is only suppressed, arising later when appropriate

3. punishment increases aggression because it models aggression as a way to cope

4. Also teaches discrimination

5. Can teach fear

6. If using it pair with reinforcement; notice the individual doing what you want and reward them (like the Norski Nod!)

Extending Skinner page 235

Do cognitive processes and biological constraints affect operant conditioning?

Just before his death, in his last speech, Skinner maintained that behavior is the key in psychology and not cognition. But, research has shown cognitive processes are much more important, even in conditioning.

Latent learning demonstrating knowledge only when it is needed; Tolman and Honzik did the research.

Ex. mice that explored a maze only demonstrate that they know the maze well by directly going to the food placed the previous time

Cognitive map a mental image of your surroundings; mice developed this in Tolman’s studies; he showed there is more to learning than just association; cognition is also at work.

Insight learning, 236

The solution suddenly appears in the mind.

Intrinsic motivation 236

Performing an activity for its own sake

Edward Deci showed that rewards can undermine motivation: if you already like doing something and are then rewarded for the activity, you may lose interest and only perform it for the reward.

Extrinsic motivation behaving in a certain way to receive external rewards

Overjustification effect giving a reward for something the organism already likes to do; here, the extrinsic motivator overrides intrinsic interest in the task

This concept is related to cognitive dissonance (Leon Festinger’s concept on page 648): in change the difference between an attitude and action people will often change their attitude to support their action.

This is unfavorable because the organism will lose interest in the activity

Ex. Paying someone to do a puzzle they already like doing

Some researchers claim rewards can raise performance and spark creativity. Don’t use them as bribes or to control, but to signal a good performance.

Biological predispositions, 237

Skinner’s students Keller Breland and Marian Breland studied animals. They can be “taught” to do certain human behaviors using operant conditioning but then revert to the normal ways; this is known as “instinctive drift” with trained animals acting naturally even though trained.

Superstitions are behaviors that are accidentally reinforced.

Practical applications of Operant Conditioning: Skinner’s legacy. Page 238

1. Improving Student performance: computers replace teachers to properly reinforce learned behavior

2.Improving Work productivity

Make reinforcement immediate

Well-defined goals that are achievable

3.Helping shape child behaviors

Reinforce good behavior

Ignore whining

Explain misbehaviors and give time outs

4.Improving athletic skills

5 Self-improvement: specify goals, monitor, reinforce, reduce rewards. Neil Miller’s concept of biofeedback can be used here. This is a cognitive strategy to monitoring of changes in physiological responses.

Contrasting Classical and Operant Conditioning

See table 6.4, page 241

III. Learning by observation ALBERT BANDURA page 242

What is observational learning and how is it enabled by mirror neurons?

Observational learning---also called social learning--- researched by Albert Bandura in the 1960s to the present.

Modeling watching specific behaviors of others and imitating them; we tend to mimic models who are similar to us.

His famous experiment was the Bobo Doll experiment. Children acted violently after watching television programming which depicts violent characters that acted as models.

Self-efficacy and learning

1. Albert Bandura: Your belief about your ability to learn to make changes in your life.

2. People high in self efficacy, have more control over their lives.

3. They take information from the environment and use it to improve their lives

4. Summary: If you believe you can do learn to do something, you are more likely to be able to achieve it.

Mirrors in the brain, 243

Mirror neurons frontal lobe neurons; they fire when performing certain actions or watching others; allows for imitation, language learning and empathy; discovery of Giacomo Rizzolatti.

We have a mirror neuron system that supports empathy (emotional understanding of others) and imitation.

These neurons help infer the mental states of other and give rise to a theory of mind. People with autism lack this ability (see 422-423)

Mirror neurons underlie our social behavior.

Bandura’s experiments page 244

Bandura was looking at vicarious learning: when you see someone else rewarded for emitting a behavior.

When the adult model was reinforced for aggressive behavior, the children who saw this in one version of the Bobo Doll experiment were far more likely to act aggressively themselves. This was in comparison to other groups of kids who saw the adult punished in another ending, or saw no consequence in this a third version. The effects were long lasting.

Applications

Behavior modeling used by businesses to train employees

Positive Observational learning

Prosocial behaviors actions that are beneficial, constructive, and not violent.

These behaviors can prompts similar behaviors in others thus, prosocial.

Research experiments show that children often do what their models (usually parents) do.

Hypocritical parents: saying one thing and doing another. Their children will say what they say and do what they do.

Antisocial effects, 247

Anti-social models can have anti-social effects

Abusive parents might have aggressive children

Television and observational learning

George Gerbner’s studies: television does not reflect the real world

Correlational studies link violence viewing with violent behavior (Craig Anderson’s research); but correlation does not mean causation.

Leonard Eron’s studies: the more hours children watch, the more they are at risk for aggression and crime

Myers asserts that he consensus is that violence on TV leads to aggressive behavior among children and teens. But this study is from 1982. The Supreme Court, in 2011, found a much more mixed picture in rejecting this claim.

Myers cites two factors:

One factor is imitation: kids copy what they see on TV.

Exposure to violence desensitizes the viewer: watching cruelty promotes indifference

Key terms: 251-252 (see study guide)

AP Quiz: 252-253

................
................

In order to avoid copyright disputes, this page is only a partial summary.

Google Online Preview   Download