Learning: Modules 20–22 — Basic Concepts, Classical Conditioning, Operant Conditioning, and Biology/Cognition

MODULE 20 | BASIC LEARNING CONCEPTS AND CLASSICAL CONDITIONING

  • Overarching idea: Learning is the process by which organisms adapt to their environments by acquiring enduring information or behaviors. Major forms covered here:

    • Classical conditioning: associative learning between two stimuli.

    • Operant conditioning: associative learning between a behavior and its consequences.

    • Cognitive learning: learning that involves mental processes such as observation and language.

  • Key definitions (general):

    • learning: the process of acquiring through experience new and relatively enduring information or behaviors.

    • associative learning: learning that certain events occur together; can involve two stimuli (classical conditioning) or a response and its consequence (operant conditioning).

    • stimulus: any event or situation that evokes a response.

    • respondent behavior: automatic responses to a stimulus.

    • operant behavior: behavior that operates on the environment to produce consequences.

  • Why we study learning: it helps us adapt to changing environments, build helpful habits, inform therapies, and understand motivation and personality.

  • Basic human learning tendencies:

    • We learn by association (Locke, Hume, Aristotle): parallel between seeing/hearing a cue and expecting a consequence.

    • Associations can operate subtly to shape habits (e.g., red pen vs black pen affecting error detection; voting place near a school influencing tax support).

    • Learning gives us hope: what is learnable may be taught; what has been learned can be changed through new learning (counseling, psychotherapy, rehabilitation).

  • Two major forms of associative learning (overview):

    • Classical conditioning: two stimuli are paired to produce a response that anticipates the second stimulus.

    • Operant conditioning: a behavior and its consequences shape future likelihood of that behavior.

  • First, we focus on classical conditioning (Pavlov, Watson era).

  • Classical conditioning: basic concepts and vocabulary

    • Neutral Stimulus (NS): a stimulus that initially elicits no specific response.

    • Unconditioned Stimulus (US): a stimulus that naturally and automatically triggers a response.

    • Unconditioned Response (UR): the natural, unlearned response to the US.

    • Conditioned Stimulus (CS): an originally neutral stimulus that, after association with the US, comes to trigger a conditioned response.

    • Conditioned Response (CR): a learned response to the previously neutral (now conditioned) stimulus.

  • Pavlov’s key insight: by repeatedly pairing a neutral stimulus (e.g., a tone) with an unconditioned stimulus (e.g., food in the mouth), the neutral stimulus becomes a conditioned stimulus that elicits a conditioned response (salivation) on its own.

  • Forms and processes in classical conditioning

    • Acquisition: the initial learning stage where the NS becomes associated with the US so that the CS elicits the CR. Typical timing: the NS (now CS) is presented just before the US; interval is often about half a second. If the US occurs before the NS, conditioning is unlikely.

    • Extinction: the diminished CR when the CS is presented alone without the US.

    • Spontaneous recovery: the reappearance of a weakened CR after a rest period.

    • Generalization: after a CS has been conditioned, stimuli similar to the CS may also elicit a CR.

    • Discrimination: learning to distinguish between the CS and other stimuli that do not signal the US.

    • Higher-order conditioning (second-order conditioning): a new NS can become a CS if it is paired with an already-conditioned stimulus (e.g., light → tone → food; light alone may elicit the CR, though typically weaker).

  • Pavlov’s classic procedures and diagrams

    • Basic schematic (before conditioning): NS, US, UR; no CR.

    • During conditioning: NS paired with US, leading to CS and CR after conditioning.

    • After conditioning: CS alone elicits CR.

    • Common notation: NS + US → CS → CR; US → UR; CS → CR.

    • General timeline often depicted as: NS pairs with US to become CS; CS then evokes CR.

  • Important example types and implications

    • Whatever the animal learns helps it survive: anticipating food, avoiding danger, locating mates, etc.

    • Biological conditioning examples across species: sea slug Aplysia (gill withdrawal) and aquarium seal (slapping for herring) illustrate associative learning in simple nervous systems and complex animals.

    • Classical conditioning is a basic form of learning that adapts organisms to their environments; it can be studied objectively through measurable responses (e.g., salivation).

  • Ivan Pavlov and John B. Watson legacy

    • Pavlov: demonstrated conditioning with dogs; foundational to understanding associative learning.

    • Watson: championed behaviorism—psychology should focus on observable behavior and laws of learning, not inner mental states. He argued that basic laws of learning are the same across species.

    • Both viewed conditioning as a universal mechanism across organisms; modern psychology recognizes additional cognitive and biological factors but still uses classical conditioning as a core framework.

  • Early classic experiments and findings

    • Pavlov’s device and procedure (tone as NS, food US, salivation UR, tone becomes CS after conditioning).

    • Four major conditioning processes observed by Pavlov: acquisition, extinction, generalization, discrimination; later work added spontaneous recovery and higher-order conditioning.

    • Watson & Rayner’s “Little Albert”: classical conditioning of fear (US = loud noise; UR = fear; CS = white rat; CR = fear of rat; later generalization to rabbit, dog, sealskin coat).

    • The ethical limits of such experiments are recognized today; they catalyzed thinking about extinction and therapeutic conditioning.

  • Applications of classical conditioning to human health and behavior

    • Drug cravings: people in drug-using contexts may experience cravings when exposed to cues associated with past highs; cue exposure therapy attempts to extinguish these associations or reframe them.

    • Food cravings: repeated associations between tastes and pleasure can make dieting challenging; conditioning can create cravings for unhealthy foods.

    • Immune responses: tastes paired with immune-modulating drugs can come to evoke immune responses on their own (taste-immune conditioning).

    • Conditioning gone wrong or used therapeutically: Pavlovian processes used to understand emotional responses and fears; Watson’s fears led to consideration of extinction and counterconditioning as therapy.

  • Conditioning and cognitive factors

    • While classical conditioning emphasizes automatic associations, cognitive processes influence the strength and speed of conditioning. Examples:

    • Predictability: Rescorla-Wagner and related research show that the more predictable an association, the stronger the CR.

    • Expectations and cognitive maps influence conditioning outcomes.

    • Preparedness and taste aversion as a clear cognitive-biological constraint: organisms are biologically predisposed to learn certain associations (e.g., taste with nausea) more readily than others (e.g., sight or sound with nausea) due to survival value.

    • Garcia & Koelling (1966) taste-aversion experiments showed that rats learned taste–illness associations readily but not sight/sound–illness associations, even when illness occurred hours after tasting. This undermined the idea that any perceptible stimulus could serve as a CS for any US.

  • Taste aversion and real-world implications

    • Taste aversion learning occurs when a taste is linked to nausea or illness; even if illness occurs hours after tasting, the taste can become a CS for nausea.

    • Preparedness: some associations are easier to learn because they are naturally adaptive (e.g., taste aversion in food safety).

    • Implications for chemotherapy patients (cancer treatment contexts): nausea can generalize to clinic cues (waiting room, nurses, etc.).

  • Connections to other learning concepts

    • Higher-order conditioning shows how CSs can acquire salience without a direct US; this process, though usually weaker, can shape everyday responses.

    • Generalization vs discrimination demonstrates adaptive flexibility: broad generalization can be useful for quick threats, while discrimination helps avoid maladaptive responses to non-threatening stimuli.

  • Recap: Why Pavlov’s work remains important

    • Demonstrated general applicability of classical conditioning across species and contexts.

    • Established objective measures of learning and a model for careful experimentation in psychology.

  • Terminology to remember (Module 20)

    • NS, US, UR, CS, CR, acquisition, extinction, spontaneous recovery, generalization, discrimination, higher-order conditioning, neutral stimulus, conditioned stimulus, conditioned response, unconditioned stimulus, unconditioned response, respondent behavior, conditioning.

  • Quick retrieval practice prompts (selected from the module)

    • RP-1: What is the NS in the dog tone–food experiment? US? UR? CS? CR?

    • RP-2: Before conditioning, what are the NS and US for Pavlov’s experiment; what are UR and CR after conditioning?

    • RP-3: In Pavlov’s setup, what is acquisition and what is extinction? How do spontaneous recovery and generalization differ?

    • RP-4: What is higher-order conditioning? Provide a brief example.

  • Preview for next module (bridge to Module 21)

    • While classical conditioning concerns associations between stimuli (and automatic responses), operant conditioning concerns associations between behaviors and consequences and emphasizes voluntary, goal-directed actions.


MODULE 21 | OPERANT CONDITIONING

  • Core idea: Operant conditioning governs how we learn from the consequences of our actions. It extends conditioning to voluntary behaviors that operate on the environment to produce outcomes.

  • Classical vs. operant conditioning: quick comparison

    • Classical conditioning

    • Form: learning associations between two stimuli that we do not control.

    • Behavior: automatic, involuntary (respondent behavior).

    • Operant conditioning

    • Form: learning associations between our behavior and its consequences.

    • Behavior: voluntary, operates on the environment to produce consequences (operant behavior).

  • B. F. Skinner and operant conditioning

    • Skinner (1904–1990): pivotal figure in modern behaviorism; expanded on Thorndike’s law of effect by developing a systematic way to study behavior and reinforcement using the Skinner box (an operant chamber) where an animal’s bar pressing or key pecking yields a reinforcer.

    • Law of effect (Thorndike): rewarded behavior tends to recur; punished behavior is less likely to recur.

    • Key idea: reinforcement, not punishment alone, shapes behavior effectively. Punishment can be swift and effective but has several drawbacks.

  • Skinner box and shaping

    • Skinner box components: bar or pulley releasing food/water; recording devices to track responses.

    • Shaping: gradually guiding an animal toward a desired behavior by reinforcing successive approximations. This method helps teach complex behaviors (e.g., pigeons walking in a figure 8, playing Ping-Pong, or keeping a missile on course by pecking a screen target).

    • Importance of discriminative stimuli: cues that signal when a response will be reinforced can be learned and used to guide behavior.

  • Reinforcement versus punishment

    • Reinforcement: any event that strengthens the behavior it follows.

    • Punishment: an event that tends to decrease the behavior that it follows.

    • Types of reinforcers

    • Positive reinforcement: add a desirable stimulus to increase a behavior. Examples: petting a dog that comes when called; paying someone for work done.

    • Negative reinforcement: remove an aversive stimulus to increase a behavior. Examples: taking painkillers to end pain; fastening a seatbelt to end the loud beeping.

    • Primary vs conditioned reinforcers

    • Primary reinforcers: innately reinforcing (e.g., food, warmth).

    • Conditioned (secondary) reinforcers: gain power through association with primary reinforcers (e.g., money, good grades, a pleasant voice).

    • Negative reinforcement is not punishment; it strengthens a behavior by removing something negative.

    • Skinner’s framework also covers how reinforcement can be used ethically to encourage prosocial behavior and discourage problematic behavior.

  • Immediate versus delayed reinforcers

    • Humans can respond to delayed gratifications (e.g., paychecks, grades, trophies).

    • Delayed gratification is a key aspect of self-control and mature decision-making (e.g., Marshmallow Test literature).

    • In shaping, immediate reinforcement tends to facilitate faster learning, but delayed reinforcement requires self-control and future payoff thinking.

  • Reinforcement schedules (partial reinforcement and extinction resistance)

    • Continuous reinforcement: reinforce after every correct response. Rapid acquisition but rapid extinction when reinforcement stops.

    • Partial (intermittent) reinforcement: reinforce only some of the time, leading to slower acquisition but greater resistance to extinction.

    • Fixed-ratio (FR) schedules: reinforce after a set number of responses (e.g., buy 10 coffees, get 1 free).

    • Variable-ratio (VR) schedules: reinforce after an unpredictable number of responses (e.g., slot machines).

    • Fixed-interval (FI) schedules: reinforce the first response after a fixed time interval (e.g., weekly test with new material arriving after fixed time).

    • Variable-interval (VI) schedules: reinforce the first response after varying time intervals (unpredictable rewards).

    • Key takeaway: higher response rates tend to occur with ratio schedules (more responses per reinforcement), and greater persistence occurs under variable schedules due to unpredictability of reinforcement.

  • Practical implications and applications

    • In schools: immediate feedback and well-defined, achievable reinforcements can boost learning; shaping toward target behaviors.

    • In sports: reinforce small successes first, gradually increase difficulty to build mastery.

    • At work: align rewards with clearly defined targets; use contingent reinforcement to boost productivity.

    • In parenting: reinforce desirable behaviors with praise or privileges; avoid whining reinforcement loops; use time-outs or removal of privileges as needed instead of harsh punishment.

    • Self-improvement: set measurable goals, track progress, reinforce progress, and gradually reduce rewards.

  • Positive and negative reinforcement in context

    • Positive reinforcement example: offering praise or a reward for a desired behavior.

    • Negative reinforcement example: removing an unpleasant task once the desired behavior is exhibited (e.g., stop nagging when the child starts studying).

    • Important nuance: negative reinforcement is not punishment; it increases desired behavior by removing something aversive.

    • Punishment examples and concerns:

    • Positive punishment: administer an aversive stimulus to reduce a behavior (e.g., spray water on a barking dog).

    • Negative punishment: remove a desirable stimulus to reduce a behavior (e.g., revoke privileges).

    • Drawbacks of punishment: can create fear, discrimination among settings, and may increase aggression; does not necessarily teach a more desirable alternative; can backfire if misapplied.

  • Latent and observational learning in operant contexts

    • Observational learning is distinct but complementary; learning can occur without direct reinforcement by watching others.

    • Latent learning: learning that is not immediately expressed until there is incentive to demonstrate it (e.g., Tolman & Honzik maze study).

    • Motivation: intrinsic vs extrinsic motivation; excessive extrinsic rewards can undermine intrinsic interest in an activity.

  • Bandura and modeling: social learning theory

    • Observational learning (modeling) involves paying attention, retaining the observed behavior, and rehearsing or reproducing it.

    • Vicarious reinforcement/punishment: learning the expected consequences by watching others experience rewards or punishments.

    • Mirror neurons and theory of mind: neural mechanisms proposed to support imitation and empathy; evidence from neuroimaging and brain research supports a neural basis for understanding others’ actions and feelings; debates exist about the exact role and universality of mirror neurons.

    • Bandura’s Bobo doll paradigm showed children imitate aggression modeled by adults, demonstrating powerful observational learning with both prosocial and antisocial outcomes.

  • Prosocial and antisocial effects of observational learning

    • Prosocial modeling: observing helpful or cooperative behavior increases similar behavior in others; can be used to train employees, children, and communities; examples include Gandhi’s nonviolent action and MLK Jr.’s leadership; prosocial media content can boost helping behaviors.

    • Antisocial modeling: exposure to aggression or deceit can increase aggression and antisocial behavior in observers; media violence correlates with increases in aggression; however, causation is complex and moderated by cognition, context, and individual differences.

    • Important caveat: media effects research often shows correlations and experimental evidence under controlled conditions; real-world outcomes depend on multiple interacting factors.

  • Cognitive and social neuroscience perspectives on observational learning

    • Observational learning engages cognitive representations and anticipatory expectations; identical brain networks may activate when observing and performing actions, supporting the idea of shared representations.

    • Empathy and theory of mind are linked to observational learning via brain systems that simulate others’ experiences, including pain and emotion.

  • Violence viewing effect: thinking critically about media violence

    • The violence-viewing effect refers to increased aggression or aggressive thoughts/affect after exposure to media violence, particularly under certain conditions (attractive aggressor, justified/unpunished violence, realistic harm).

    • Evidence suggests a pattern of increased aggressive behavior, aggressive cognitions, and decreased prosocial behavior, but the magnitude and universality vary; causation is multifactorial and contextual.

    • APA Task Force and various researchers summarize a nuanced view: media violence is a risk factor for aggression, not a sole cause.

  • Key terms and connections (Module 22)

    • Observational learning, modeling, mirror neurons, cognitive map, latent learning, intrinsic motivation, extrinsic motivation, prosocial modeling, antisocial modeling.

    • Biological and cognitive influences on conditioning (Table 22.1): organisms are influenced by natural predispositions and cognitive expectations; preparedness shapes what associations are readily learned; latent learning and cognitive maps illustrate mental representations beyond simple stimulus–response links.

  • Applications of observational learning to real-world contexts

    • Prosocial modeling can improve social behavior in workplaces and families; modeling effective communication and cooperative skills.

    • Antisocial modeling underscores the importance of media literacy and parental modeling to reduce aggression and aggression-prone behaviors.

    • Observational learning explains language acquisition, moral development, and cultural transmission through social modeling.

  • Think critically about: The effects of viewing media violence (summary)

    • Media violence exposure is associated with increased aggression in some studies; causal links are often mediated by cognitive, social, and personal factors; careful interpretation of correlations and experimental data is necessary.

  • Summary of Module 22

    • Conditioning is not solely a product of stimulus–response; biology, cognition, and social context collectively shape learning.

    • Preparedness and instinctive drift illustrate how biology constrains learning and shapes what is easy to condition.

    • Observational learning, cognitive processes, and mirror-neuron research provide a richer understanding of how humans learn from others, including prosocial and antisocial outcomes.


Connections Across Modules (20–22)

  • Classical conditioning (Module 20) provides the foundation for understanding how associations form between stimuli. Biological constraints (Module 22) explain why some associations are easier to learn than others and why certain aversions (e.g., taste aversion) are particularly robust.

  • Operant conditioning (Module 21) extends learning to voluntary behaviors and consequences, with practical implications for education, workplace productivity, parenting, and self-improvement. It also introduces the idea that reinforcement schedules and the immediacy of reinforcement influence how quickly and how persistently behaviors are learned.

  • Observational learning (Module 22) integrates cognition and social context by showing how we learn from watching others, including the role of cognitive representations, imitation, and neural mechanisms (mirror neurons).


Key Numbers, Concepts, and Equations (LaTeX)

  • Classical conditioning core relationships:

    • Before conditioning: NS → no response; US → UR.

    • After conditioning: CS → CR.

    • Acquisition timing: NS + US with a short interval (often ≈ 0.5 seconds).

    • Higher-order conditioning: CS1 → CS2 → US; CS2 → CR (often weaker).

  • Habit formation data:

    • Habit formation averages: about 66 days for a behavior to become automatic, based on a study with 96 university students over 84 days (behavior felt automatic after about 66 days).

  • Shaping and reinforcement examples: immediate reinforcement tends to produce faster learning; delays can impede acquisition if the reinforcer is delayed beyond roughly 30 seconds for animals in shaping scenarios.

  • Reinforcement schedules (conceptual notation)

    • Continuous reinforcement: every correct response is reinforced.

    • Fixed-ratio (FR) schedule: every nth response is reinforced.

    • Variable-ratio (VR) schedule: reinforced after an unpredictable number of responses.

    • Fixed-interval (FI) schedule: reinforcement after a fixed amount of time has elapsed.

    • Variable-interval (VI) schedule: reinforcement at unpredictable time intervals.

  • Taste aversion and preparedness (biological constraints):

    • Taste aversion learning often occurs when a taste (CS) is followed by illness (US) with a delay, which is adaptive for survival; not all stimuli are equally readily associated with a US.

  • Mechanisms and measures:

    • Response rates under different reinforcement schedules show higher rates with ratio-based reinforcement, and greater persistence with variable/unpredictable schedules.


Quick Retrieval Practice (mini-prompts)

  • From classical conditioning: What is the NS in Pavlov’s dog experiment? What is the US? UR? CS? CR?

  • In operant conditioning, differentiate positive reinforcement, negative reinforcement, positive punishment, and negative punishment with an example each.

  • What is shaping, and how does it differ from simple reinforcement?

  • How does a fixed-interval schedule typically shape responding over time?

  • What is observational learning, and how do mirror neurons relate to imitation?

  • How can biology constrain conditioning (preparedness and instinctive drift)? Give an example.

  • What is the violence-viewing effect, and what factors influence its strength?

  • Define latent learning and provide an example (maze-learning with/without rewards).

  • Connections to real-world scenarios:

    • Design a classroom reinforcement plan using a fixed-ratio or variable-ratio schedule to promote consistent studying.

    • Develop a prosocial modeling intervention for a family or classroom that emphasizes positive imitation and vicarious reinforcement.


Terminology to Remember (brief recap across Modules 20–22)

  • Classical conditioning: linking two stimuli; involuntary responses.

  • Operant conditioning: linking behavior with consequences; voluntary responses.

  • Acquisition, extinction, spontaneous recovery, generalization, discrimination.

  • NS, US, UR, CS, CR; reinforcement; punishment; positive/negative reinforcement; primary/conditioned reinforcers.

  • Shaping, discriminative stimulus.

  • Latent learning, cognitive map.

  • Observational learning, modeling, vicarious reinforcement/punishment.

  • Biological constraints, preparedness, instinctive drift, taste aversion.

  • Mirror neurons, theory of mind, prosocial/antisocial modeling.

  • Violence-viewing effect; cognition’s role in conditioning.

This set of notes covers the major and many of the minor points discussed across Modules 20–22, including the learning forms, core experiments, key processes, real-world applications, and the cognitive and biological factors that shape learning. Use the retrieval prompts and diagrams from the text to reinforce understanding and to prepare for exam-style questions.