Behaviorism and Social Cognitive Theory

Behaviorism: origins, scope, and core assumptions

  • Behaviorism emerged in the early 1900s as a response to Freud's emphasis on unconscious processes and inner drives. North American psychology pushed for a more scientific, environment-focused approach and was skeptical of unconscious explanations.

  • Core idea: psychology should focus on environmental forces acting on behavior; thoughts and feelings are either ignored or treated as black boxes because they are not directly observable.

  • Key stance on personality (from behaviorists): personality is a set of learned responses to environmental stimuli; stable individual differences are reinterpreted as patterns of learned behavior rather than deep inner traits. Internal mental states are acknowledged but not the primary subject of study.

  • Fundamental concepts:

    • Learning is a relatively stable change in behavior resulting from experience.

    • Behaviorism emphasizes observable stimulus–response (S–R) relations and often uses animals (e.g., pigeons, rats) to infer universal laws.

    • Thoughts and feelings are either not studied directly or are inferred only through external manifestations (e.g., crying as an index of sadness).

  • Behaviorism ultimately excluded many aspects of personality as they argued those aspects could not be directly measured, which led to criticisms and a shift toward more integrative theories later.

Classical conditioning (Pavlovian conditioning)

  • Classical conditioning is a type of associative learning where a neutral stimulus becomes associated with a meaningful stimulus, eliciting a learned response.

  • Core terms and notation:

    • Unconditioned stimulus (US): elicits an automatic response without learning. E.g., food causing salivation.

    • Unconditioned response (UR): natural, unlearned reaction to the US.

    • Neutral stimulus (NS): initially elicits no response of interest (e.g., a bell).

    • Conditioned stimulus (CS): originally NS that, after pairing with US, triggers a learned response.

    • Conditioned response (CR): learned response to the CS, often resembling the UR.

  • Classical conditioning sequence (typical example: Pavlov’s dogs):

    • US (food) → UR (salivation)

    • NS (bell) paired with US

    • After conditioning, CS (bell) → CR (salivation)

  • Basic principles:

    • Acquisition: gradual association of CS with US leading to CR.

    • Extinction: weakening of CS–CR association when CS is presented without US.

    • Spontaneous recovery: reappearance of a lost CR after a rest period.

    • Generalization: CR occurs to stimuli similar to CS.

    • Discrimination: learning to respond only to the CS and not to similar stimuli.

  • Historical exemplars and details:

    • Ivan Pavlov (physiologist, not primarily a psychologist) demonstrated that dogs salivate to a neutral stimulus that predicts food (e.g., a bell).

    • John B. Watson popularized behaviorism and applied classical conditioning to emotions (e.g., Little Albert experiment).

    • Little Albert: nine-month-old infant conditioned to fear a white rat after pairing the rat with a loud noise; fear generalized to other white, fluffy objects. This study highlighted ethical issues in psychological research.

    • Mary Cover Jones (mother of behavior therapy) developed counterconditioning to reduce fears by pairing feared stimuli with positive experiences (e.g., Peter the boy’s fear of white rabbits reduced by giving candy during gradual exposure).

    • Systematic desensitization (Joseph Wolpe, sometimes cited as Joseph Wolf in some notes): staged exposure to feared stimuli paired with relaxation techniques (e.g., vagal breathing) to reduce anxiety; used a hierarchy of feared situations.

  • Practical applications and examples:

    • Advertising and consumer behavior: classical conditioning used to pair products with positive images or feelings (e.g., toothpaste with attractive imagery) to create positive associations.

    • Sniffer bees: trained bees learn to associate explosive smells with sugar rewards; after conditioning, bees respond to explosive odors with the proboscis (sugar-detection reflex). This demonstrates how conditioning can be used beyond humans.

    • VR-based exposure therapy: using virtual reality to provide graded exposure (e.g., sharks) while practicing relaxation techniques to reduce phobias (systematic desensitization in a modern format).

    • Everyday phenomena: phobias and conditioned emotional responses; counterconditioning shows how maladaptive responses can be reversed.

  • Conceptual takeaways:

    • Conditioning can explain many reflexive and emotional responses as products of environmental pairings.

    • Early behaviorists often treated emotion, cognition, and personality as reducible to learned associations, sometimes overlooking internal cognitive processes.

Nonassociative learning (habituation and sensitization)

  • Nonassociative learning involves changes in response to a single stimulus without forming new stimulus–stimulus connections.

  • Habituation: gradual decrease in response to a repeated, benign stimulus (e.g., smell of strong cheese becoming less noticeable after exposure).

  • Sensitization: increase in response to a stimulus following noxious or intense exposure (e.g., worsening irritation from a repeated cough during a lecture).

  • Distinction: habituation reduces response to the same stimulus; sensitization increases responsiveness to the stimulus.

Operant conditioning (instrumental conditioning)

  • Definition: a form of learning in which the consequences of a behavior determine its likelihood of reoccurrence.

  • Key features:

    • Focus on voluntary behaviors (as opposed to involuntary reflexes in classical conditioning).

    • Behavior can be spontaneously emitted and shaped over time by consequences.

    • The environment shapes behavior through reinforcement and punishment.

  • Distinctions from classical conditioning:

    • Classical conditioning links a neutral stimulus to an involuntary response; operant conditioning links a behavior to its consequences.

    • In operant conditioning, what comes after the behavior (consequence) determines future likelihood, not what happens before.

  • B. F. Skinner and the experimental apparatus:

    • Skinner boxes (operant conditioning chambers) with rats or pigeons to study how behaviors are shaped by consequences.

    • Bar presses or pecking lead to rewards (food) or punishments (aversive stimuli).

  • Core concepts:

    • Reinforcement: any consequence that increases the likelihood of a behavior.

    • Primary reinforcers: satisfy innate biological needs (e.g., food, water, sleep, shelter, sex, affection).

    • Secondary reinforcers: learned reinforcers that acquire value by associated access to primary reinforcers (e.g., money, praise, good grades).

    • Punishment: any consequence that decreases the likelihood of a behavior.

    • Positive punishment: adding something unpleasant (e.g., loud noise).

    • Negative punishment: removing something desirable (e.g., loss of privileges).

    • Four main types of consequences:

    • Positive reinforcement: add a pleasant outcome to increase behavior.

    • Negative reinforcement: remove an aversive outcome to increase behavior.

    • Positive punishment: add an aversive outcome to decrease behavior.

    • Negative punishment: remove a desirable outcome to decrease behavior.

  • Schedules of reinforcement (how often a behavior is reinforced):

    • Continuous reinforcement (CRF): reinforcement after every correct response; fast learning but high extinction when reinforcement stops.

    • Intermittent reinforcement (partial): reinforcement only some of the time.

    • Fixed interval (FI): reinforced after a fixed amount of time (e.g., weekly paycheck).

    • Variable interval (VI): reinforcement after varying time intervals.

    • Fixed ratio (FR): reinforcement after a fixed number of responses (e.g., buy 10 coffees, get 1 free).

    • Variable ratio (VR): reinforcement after a variable number of responses (high resistance to extinction; often most addictive in behavior).

  • Shaping: gradually reinforcing closer approximations toward a target behavior.

    • Process: reinforce any behavior that resembles the desired behavior, then gradually require more precise forms before reinforcement; shapes complex actions (e.g., teaching a dog to put all four paws in a box or a child to shade an image progressively).

  • Important cautions and applications:

    • Punishment can suppress behavior but often does not teach the desired alternative behavior; reinforcement is generally more effective for long-term behavior change.

    • Prolonged or inconsistent punishment can lead to fear, aggression, or avoidance rather than learning.

    • Real-world examples: superstition (unrelated consequences become linked to a behavior), gambling (variable-ratio reinforcement creates strong, resistant-to-extinction behaviors), love bombing (initial continuous reinforcement followed by intermittent reinforcement to maintain control).

  • Notable historical figures and notes:

    • B. F. Skinner: radical behaviorist; advocated reinforcement over punishment; emphasized operant conditioning and behavior modification.

    • Skinner boxes facilitated the study of how reinforcement schedules shape behavior and how punishment can be used (though not as effectively as reinforcement).

    • Skinner’s broader impact extended to education, childcare, and organizational settings; cautions about misusing punishment in schools and parenting.

  • Real-world implications:

    • Conditioning principles underlie advertising, habit formation, and addictive behaviors (e.g., gambling and video games using variable ratio reinforcement).

    • Understanding shaping helps explain how complex skills can be taught step-by-step (e.g., training animals or teaching tasks in education).

Classical conditioning applications in everyday life

  • Advertising and consumer psychology:

    • Repeated pairings of a product (neutral stimulus) with a desirable image or celebrity (unconditioned or positive stimulus) create positive associations with the product (toothpaste example).

    • The goal is to elicit positive affect toward the product without relying solely on its stated functional benefits.

  • Beekeeping and odor detection (sniffer bees):

    • Bees naturally extend their tongues when they smell sugar; after conditioning, the odor of explosives triggers the same reflex, enabling explosion detection.

    • Advantages: bees can be trained rapidly (5–6 rounds) and can outperform dogs on some smell tasks; ongoing applications include drugs, disease detection, and cancer biomarkers from breath.

  • Aversion therapy and exposure through systematic desensitization:

    • Systematic desensitization blends relaxation techniques (e.g., diaphragmatic or vagal breathing) with controlled exposure to anxiety-evoking stimuli along an anxiety hierarchy, gradually reducing fear responses.

  • Phobia treatment through counterconditioning (Mary Cover Jones):

    • Replacing an undesired response (fear) with a desired one (calm or positive response) by conditioning a competing response to reduce fear.

  • Case study highlights:

    • Sharks phobia treated via immersive VR exposure with relaxation training; gradual exposure reduced fear so the patient could swim more confidently.

Observational learning and social cognitive theory

  • Observational learning (modeling) is learning that occurs by watching others’ behavior and outcomes and then modeling those behaviors.

  • Not limited to humans; other social species (e.g., orangutans) show observational learning in complex tasks.

  • Distinguishing imitation from genuine observational learning:

    • Mindless imitation: reflexive, automatic copying (yawning, etc.).

    • Observational learning (model-based learning): involves cognitive processing to map observed behaviors and adapt them to new contexts.

  • Factors influencing what we model:

    • Characteristics of the model: prestige, similarity, attractiveness, and power increase the likelihood of being modeled.

    • Observer attributes: high motivation, self-esteem, and dependence on others influence likelihood of modeling.

  • Mediational processes (Bandura): the cognitive processes between observing a model and reproducing the behavior. Four key processes:

    • Attention: must attend to the model; with the internet and social media, geographic boundaries matter less for selecting models.

    • Retention: remember observed behavior via mental imagery or verbal encoding.

    • Motor reproduction: translate memories into action; requires practice, especially for complex behaviors.

    • Motivation: willingness to perform the observed behavior, influenced by anticipated rewards or punishments (vicarious reinforcement/punishment).

  • Vicarious reinforcement and punishment: learning can occur by observing others’ outcomes, not just direct consequences to oneself.

  • Albert Bandura and the Bobo doll experiments:

    • Children observe an adult behaving aggressively toward a Bobo doll (verbal aggression and physical aggression).

    • Children exposed to aggressive models show increased aggression themselves, especially in the same specific behaviors demonstrated by the model.

    • Findings generalize: exposure to aggressive models increases aggression in children.

  • Contemporary research and debates:

    • Xiao and Wang (2019) explored observational learning and aggression via violent video games; found that higher aggression beliefs may be fostered by violent games, but the effect is moderated by family environment.

    • Positive family environments can buffer or mitigate aggression even when violent video games are frequently played, suggesting social context matters.

  • Julian Rotter and social cognitive theory—locus of control and expectancy-value dynamics:

    • Behavior potential (BP) formula: BP = E × RV, where

    • E (expectancy) is the perceived probability that a given behavior will lead to a desired outcome.

    • RV (reinforcement value) is how desirable the outcome is to the individual.

    • Expectancies are influenced by prior experiences and internal factors such as locus of control and motivational states.

    • Reinforcement value reflects how much the outcome is valued by the person; both factors influence the likelihood of engaging in a behavior.

Self-efficacy and Bandura’s social cognitive theory

  • Self-efficacy: belief in one’s ability to organize and execute the actions required to manage prospective situations.

    • Distinct from general personality traits; self-efficacy is variable across domains (academic, health, coping, etc.) and can change with context.

    • High self-efficacy leads to greater motivation to tackle challenging tasks, better cognitive control, reduced anxiety, and perseverance in the face of difficulty.

    • Conversely, low self-efficacy can lead to avoidance, rumination, and reduced performance expectations.

  • Four sources of self-efficacy:

    • Master experiences (past successes build belief in one’s abilities).

    • Modeling (observing similar others succeed increases belief in one’s own ability).

    • Social persuasory support and physiological states (positive social feedback; lower cortisol; better emotional regulation).

    • Psychological factors (perception of emotions, stress responses, and cognitive interpretations of arousal).

  • Practical implications and strategies to boost self-efficacy:

    • Break big goals into smaller, manageable subtasks to create mastery experiences.

    • Use appropriate role models similar to the individual to enhance modeling effects.

    • Employ cognitive strategies and self-talk to regulate anxiety (e.g., third-person self-talk using one’s own name to create cognitive distance and reduce anxiety).

    • Reframe anxiety as a signal of meaningful engagement and motivate action rather than suppressing arousal.

  • Effects and benefits of high self-efficacy:

    • Improved performance across domains, better coping with stress, and better adaptive behaviors.

    • Positive feedback cycles: higher self-efficacy leads to better performance, which further reinforces self-efficacy.

  • Agency and human potential:

    • Bandura emphasized human agency: people are not just passive outcomes of their environments; they influence outcomes and environments through intentional action and self-regulation.

    • The concept of deliberate forethought, intentionality, self-regulation, and self-reflection underpins agency in social cognitive theory.

Reciprocal determinism and human agency in Bandura’s framework

  • Reciprocal determinism: behavior, personal factors (cognition, emotions), and environmental factors all influence each other.

  • Individuals are not merely products of their environments; they also shape their environments and their own experiences through choices and actions.

  • Observational learning and agency operate together: people select, attend to, and interpret social information and adjust their behavior accordingly.

Locus of control (Rotter) and expectancy-value thinking

  • Internal vs external locus of control:

    • Internal locus of control: outcomes are a function of one’s own efforts and abilities.

    • External locus of control: outcomes are due to luck, fate, or powerful others.

  • Impact of locus of control on behavior and life outcomes:

    • Higher internal locus of control is associated with better academic performance, healthier behaviors, and more proactive coping.

    • Higher external locus of control is associated with greater susceptibility to external influences and less adaptive learning from mistakes.

  • Assessment and practical use:

    • Rotter developed a locus of control scale (widely used in psychology) to assess individuals’ belief systems about control over events.

  • Local control measures and therapeutic implications:

    • Clinicians use locus of control concepts to tailor interventions and to empower clients to take more active control over health, education, and life outcomes.

Real-world implications and cautions

  • Advertising and media: conditioning principles underpin strategies to create positive associations with products and brands; the goal is to influence affective responses, attitudes, and consumer behavior.

  • Gambling and gaming: variable ratio reinforcement is a particularly potent schedule that sustains engagement and can contribute to addiction-like patterns.

  • Relationships and manipulation: love bombing illustrates how initial constant reinforcement followed by intermittent reinforcement can foster dependency and control.

  • Ethics in research: classic studies (e.g., Little Albert) raised ethical concerns about inducing fear or distress in participants, highlighting the evolution of ethical standards in psychology.

Summary of key concepts and references

  • Classical conditioning: CS, US, UR, CR; acquisition; extinction; generalization; discrimination; higher-order conditioning.

  • Operant conditioning: reinforcement types (positive/negative; primary/secondary reinforcers), punishment types (positive/negative), shaping, and schedules (CRF, FI, VI, FR, VR).

  • Observational learning: attention, retention, production, motivation; model characteristics; vicarious reinforcement/punishment; mediational processes.

  • Self-efficacy: sources (mastery experiences, modeling, physiology, social persuasions); domain-specific self-efficacy; impact on motivation and performance.

  • Locus of control (Rotter): internal vs external; implications for learning and behavior.

  • Behavior potential (Ladder/Rotter): BP = E × RV, with expectancy (E) and reinforcement value (RV) as core determinants of behavior

  • Real-world examples and cautions: advertising, sniffer bees, systematic desensitization and VR exposure, shaping in education, aggression studies, and the role of family context in media effects.

Quick reference formulas and terms (LaTeX)

  • Classical conditioning sequence:

    • US \rightarrow UR

    • NS + US \rightarrow UR

    • CS \rightarrow CR

  • Behavior potential (Rotter):

    • BP = E \times RV

    • where E = expectancy (belief that a given behavior will lead to a desired outcome) and RV = reinforcement value (desirability of the outcome).

  • Reinforcement types (summary):

    • Positive reinforcement: add a positive consequence to increase behavior.

    • Negative reinforcement: remove an aversive stimulus to increase behavior.

    • Positive punishment: add an aversive stimulus to decrease behavior.

    • Negative punishment: remove a desirable stimulus to decrease behavior.

  • Schedules of reinforcement (typical categories):

    • Continuous reinforcement (CRF)

    • Fixed interval (FI)

    • Variable interval (VI)

    • Fixed ratio (FR)

    • Variable ratio (VR)

  • Key components in observational learning (Bandura): attention, retention, production (motor reproduction), motivation (vicarious reinforcement/punishment).

  • Core terms: NS, CS, US, UR, CR; acquisition; extinction; generalization; discrimination; shaping; reciprocity of determinism; self-efficacy; agency; locus of control.