Avoidance & Punishment
The lecture on Avoidance and Punishment covers several key concepts:
Avoidance
Avoidance Learning: This involves learning to prevent an aversive event before it occurs. It can be either active (taking action to avoid the event) or passive (refraining from an action to avoid the event)
Theories of Avoidance:
Two-Factor Theory: This theory combines classical and operant conditioning. It suggests that avoidance is driven by an escape from fear rather than the prevention of the aversive event
Cognitive Theory of Avoidance: This theory posits that avoidance behavior is based on the expectation that a response will prevent an aversive stimulus, rather than on fear
Punishment
Positive Punishment: Involves adding an aversive stimulus to decrease a behavior. For example, scolding a child for misbehavior
Negative Punishment: Involves removing a desirable stimulus to decrease a behavior. For example, taking away a child's toy for misbehavior
Avoidance Account of Punishment: Punishment does not directly weaken a behavior but involves avoidance learning, where the avoidance response consists of any behavior other than the behavior being punished. This can lead to behaviors such as becoming a better liar or learning where speed cameras are located
Key Points
Aversive Control: Positive punishment or aversive events motivate escape behaviors, which are then strengthened via negative reinforcement
Expectations and Learning: Control over an aversive stimulus allows subjects to form expectations about their ability to influence outcomes, aligning with cognitive theories that emphasize perceived control in learning
Week 13: Avoidance and Punishment
Avoidance
Q: What is the difference between Escape behaviour and avoidance behaviour?
A: Escape behaviour involves getting away from an aversive stimulus that is in progress, while avoidance behaviour involves stay/getting away from an anticipated aversive stimulus
Q: How is avoidance often learned and how is it measured?
A: Avoidance behaviour is learned by acquiring associations between some cue and an aversive stimulus and then acquiring a response that mitigates the aversive stimulus. To measure avoidance learning, we measure the time between the presentation of the cue and the performance of the response with shorter times signifying greater learning
Q: What is the progress of learned hopelessness?
A: A learner experiences a bad event in which there is no way to control/avoid the outcome. This inability to avoid the negative outcome creates a perceived lack of control which fuels generalized helpless behaviour.
Q: Describe active and passive avoidance
A: Active avoidance involves actively making a response to avoid an aversive event, while passive avoidance involves not putting yourself in a situation where you could encounter the aversive stimulus
Q: Why are avoidance behaviours so difficult to extinguish?
A: Because the expected result of the avoidance behaviour is to not experience the aversive stimulus, if you try to extinguish the avoidance behaviour by removing the aversive stimulus the result is still the same. Thus, no extinction occurs
Theories of Avoidance
Q: Describe Mowrer’s Two Factor Theory of Avoidance
A: Avoidance learning requires two factors: 1) the learner must learn an association between and aversive stimulus and a cue, thus, causing the learner to fear the cue, 2) the learner must perform a behaviour that becomes negatively reinforced through the avoidance of the aversive stimulus
Q: What are two findings that support the Two Factor Theory?
A: The theory predicts that avoidance behaviour will be stronger if the association between aversive stimulus and the cue is stronger. Thus, factors that can weaken the association between the aversive stimulus and cue, such as adding a delay, should also weaken the avoidance behaviour. This predict was corroborated by Kamin’s (1957) which found that introducing a delay between the cue and aversive stimulus weakened avoidance behaviour. Additionally, another study found that terminating the cue immediately after the avoidance behaviour is performed strengthened avoidance behaviour
Q: What are some limitations of the Two Factor Theory?
A: The Two factor theory suggest that fear of the aversive stimulus, which is triggered by the cue, motivates avoidance behaviours, however, once the avoidance behaviour is well-learned learners no longer exhibit apparent fear. Additionally, avoidance learning can be learned without a cue, such as in Herrnstein & Hineline (1966) and the Sidman Free-Operant Avoidance experiment.
Q: In response to the above limitations, how did supporters of the Two factor theory respond? Is the response supported by evidence?
A: By devising the alternation of behaviour (yo-yo): Successful avoidance puts the CS into extinction, in which fear drops, which leads to weaken avoidance behaviours and leads to more punishment, which reinforces the avoidance behaviours again. This account is not well-supported by evidence
Q: Describe the One Factor Theory of avoidance learning
A: A behaviour that avoids an aversive stimulus is negatively reinforced (essentially the two factor theory without the first factor.
Q: What is the major limitation of the one factor model?
A: Having a cue can actually improve avoidance learning
Q: Describe the cognitive theory of avoidance learning
A: Avoidance learning is determined by expectations rather than fear. Two expectations are important: 1) the expectation that the aversive stimulus will be presented after the cue, and 2) the expectation that performing a response will lead to avoiding the aversive stimulus.
Punishment
Q: Describe the two different kinds of negative punishment
A: Timeout (the temporary removal of access to positive reinforcement) and Response Cost (Removal of a reinforcer for the problem behaviour)
Q: What is the difference between intrinsic and extrinsic punishment?
A: Intrinsic punishment is when the behaviour itself can be aversive (e.g. Hurting your back from lift heavy objects) while extrinsic punishment is when the aversive response comes from something other than the behaviour (e.g. being yelled at for lifting heavy objects without proper support/form)
Q: List the problems potential problems with using punishment
1 | Punishment of an inappropriate behaviour doesn’t reinforce an appropriate behaviour |
2 | The person delivering the punishment might become a discriminative stimulus |
3 | Given number 2, punishment might just teach the person to avoid the punisher |
4 | Punishment can elicit a strong emotional response that can inhibit learning |
5 | Punishment can elicit an aggressive response |
6 | The use of punishment might teach someone that punishment is an appropriate method for controlling behaviour |
7 | The use of punishment is often strongly reinforced |
Q: What are the benefits of using punishment
A: While rewards can help establish cooperation, without sanctions for inappropriate behaviour, cooperation may not be sustained
Q: What can be done to ensure that punishment is effective
1 | Punishment should be immediate |
2 | Punishment should be intense enough from the outset to suppress behaviour |
3 | Punishment should be consistent |
4 | Negative punishment is generally preferrable to positive punishment |
5 | Providing an explanation for why punishment is occurring can make it more effective |
6 | Punishment should also be combined with reinforcement of appropriate behaviour |
Accounts/Theories of Punishment
Q: Describe the Conditioned Suppression Account of punishment
A: Punishment doesn’t actually weaken a behaviour, but results in an emotional response that suppresses the performance of the behaviour
Q: Describe the Avoidance Account of punishment
A: Punishment actually involves avoidance learning, where the avoidance response consists of any behaviour other than the one being punished
W13: Avoidance, Punishment, and Revision.
|
Test Your Knowledge Quiz |
Q1: _____ is jumping out of a pool when you realize it's too cold while _______ is not getting in the pool at all because you checked and knew it was cold *ESCAPE: Getting out of an already bad situation) *AVOIDANCE: Preventing the bad situation entirely * avoidance and escape are learned through operant conditioning b- these terms describe consequences of behaviour, not the behaviours themselves d- these are learning mechanisms, not specific behaviours *example: ESCAPE: You smell smoke, feel heat, then run (responding to actual danger/US) AVOIDANCE: You hear fire alarm (CS), leave before any smoke/heat (US) | A1: a) avoidance, escape b) escape, avoidance c) positive reinforcement, negative reinforcement d) classical conditioning, operant conditioning
|
Q2: How does escape behaviour develop into avoidance behaviour? Nw: *CS = the warning sign *US = the unpleasant stimuli c- does not correctly account for the temporal relationship and purpose of CS as a warning signal. Correct sequence: CS → US and the goal of preventing US exposure. | A2: a) we learn experientially to act at the conditioned stimulus (CS) rather than waiting to experience the unconditioned stimulus (US) b) we learn experientially to act at unconditioned stimulus (US) rather than waiting to experience conditioned stimulus (US) c) we learn experientially to simultaneously process both CS and US, allowing for immediate response to either stimulus |
Q3: In an avoidance learning experiment, how can we tell when a subject has switched from escape behaviour to successful avoidance behaviour? When the subject: * Successful avoidance learning is demonstrated by quick, consistent responses to the warning signal (short latencies), showing the subject has learned to act before the aversive stimulus occurs. | A3: a) shows consistently short response latencies to the warning signal b) shows increasing response latencies over trials c) stops responding to the warning signal entirely d) shows variable response times between 50-100 seconds |
Q4: What type of synapse uses neurotransmitters to convey messages? *not all hormones, only dop, epinephrine/nor which serve as both hormone/nt *hormones travel in blood via endocrine system | A4: a) electrical b) chemical c) electro-chemical |
Q5: Which of the following correctly matches the brain region to its primary role in memory or learning? a- involved in nondeclarative memory, especially procedural learning and motor coordination b- primarily involved in declarative memory, such as forming and storing episodic and spatial memories. Motor learning and movement coordination involves cerebellum. | A5: a) Cerebellum: Plays a role in declarative memory, including the recollection of facts and events. b) Hippocampus: Associated with motor learning and movement coordination. c) Prefrontal cortex: Responsible for working memory, involved in planning and decision-making.
|
Q6: Which of the following schedules of reinforcement best describes a fixed interval schedule? a- reinforcement is given for the first response after a fixed amount of time has passed. b- variable ratio schedule reinforcement is based on an average number of responses, but the exact number varies. c- variable interval schedule, where reinforcement is given for the first response after a variable amount of time has elapsed, making the timing of reinforcement unpredictable. | A6: a) A rat receives a food pellet for the first lever press after a set amount of time has elapsed since the last pellet was delivered. b) A rat receives a food pellet on average every ten lever presses, but the exact number varies. c) A rat receives a food pellet after an unpredictable amount of time passes, regardless of how many times it presses the lever. |
Q7: Which non-associative learning process explains the increased ‘response’ to a drug following repeated administration? a- decrease in response, b- increase in response, more sensitive reward system = neuroadaptations = dependence. *not necessarily the effect, c- associative
*sensitisation: increased response in cravings, cues, drug seeking (behavioural/psych) *tolerance: reduced response to pleasurable effects (physical) = addiction; needs more for pleasure (tolerance) but more sensitive to cues (craving nw: tolerance is not habituation; habituation = filters non sig stimuli ) | A7: a) habituation b) sensitisation c) operant conditioning |
Q8: Which of the following NT is classified as an amino acid? a) neither an amino acid nor monoamine b) monoamines c) amino acid | A8: a) acetylcholine b) noradrenaline c) glutamate |
Q9: Electrical communication between cells occurs via ____ whereas chemical communication occurs via ____ | A9: a) gap junctions, synaptic cleft b) synaptic cleft, gap junctions |
Q10: Drugs that bind to a receptor and block or inhibit the effect of neurotransmitters are called: a- agonists: mimic or enhance the effect of neurotransmitters e.g., opioid agonist morphine b-antagonists: block or reduce the response e.g., opioid antagonist naloxone *opioids implicated w/ GABA but separate receptors | A10: a) agonists |
Q11: Which example illustrates desensitization? a- habituation, where response decreases after repeated exposure to the same stimulus. b- reducing a response to a stimulus through gradual exposure. c- sensitization, where exposure to a startling stimulus leads to an increased response to subsequent stimuli, even if they are less intense.
| A11: a) A person jumps at the sound of a loud noise but over time stops reacting as strongly. b) A person becomes less afraid of spiders after gradually being exposed to them in a controlled and progressive manner. c) After being startled by a loud noise, a person's sensitivity to sound increases, causing them to jump at even quiet noises for the rest of the day. |
Q12: Which of the following has a greater effect on dopamine? *a=large, b=some c= ‘better than expected’ | A12: a) anticipation of the reward b) the reward itself c) error in the value of the reward |
Q13: Which of the following is true? Dorsal roots carry sensory information (such as touch, pain, temperature) from the body to the spinal cord. Sensory input. AFFERENT: BODY 🡪 CNS
Ventral roots send motor info from the spinal cord to muscles. Control movement. EFFERENT: CNS 🡪BODY | A13: a) ventral roots send motor information to the spinal cord from the muscles to control movement. b) dorsal roots receive sensory information from the body and send it to the spinal cord. c) both dorsal and ventral roots receive sensory information from the body and send it to the spinal cord. |
Q14: Which scenario best represents trace conditioning? a- the conditioned stimulus (tone) and unconditioned stimulus (air puff) are presented separately with a time gap in between. b- delay conditioning, where the conditioned stimulus (light) and the unconditioned stimulus (shock) overlap. c- backward conditioning, where the unconditioned stimulus (food pellet) precedes the conditioned stimulus (light). This sequence is ineffective for forming strong associative learning because the predictor (light) does not precede the event it is supposed to predict (receiving food). | A14: a) A tone is sounded and stops before a puff of air is directed at a person’s eye, causing them to blink. b) A light is turned on, and while it remains on, a shock is administered. c) After a rat receives a food pellet, a light is turned on in its cage. |
Q15: What correctly describes the changes in membrane potential during an action potential? a- Depolarization makes the inside of the cell less negative (i.e., more +) due to the entry of Na+ ions, while repolarization restores the negative internal environment by expelling positive ions (K+) b- reverses the order of membrane potential changes during an action potential. Depolarization precedes the action potential, not follows it. c- fails to recognize that repolarization involves the cell becoming more negative, not less, as it returns to its resting state after the peak of the action potential.
Temporal sequence: depolarisation (more +), action potential, repolarisation (more -), refractory | A15: a) Depolarization occurs when the cell becomes more positive, leading to an action potential; repolarization occurs when the cell becomes more negative, returning towards the resting potential. b) Depolarization occurs when the cell becomes more negative, leading to an action potential; repolarization occurs when the cell becomes more negative, returning towards the resting potential. c) Both depolarization and repolarization result in the cell becoming less negative, leading to a sustained action potential. |
Q16: Which of the following statements about GABA and glutamate is correct?
*GABA: inhibitory, modulates processes. *Why inhibitory? inhibits the neurons ability to fire action potentials e.g., GABA binds to receptor = open ion channel, influx of Cl- *binding of NT to subunit (NT gated ion channels composed of multiple subunits)
Glutamate: excitatory, learn/mem, neuroplast, exocitocity (nerve cells are damaged/killed due to overstimulation)
| A16: a) GABA is excitatory, while glutamate is inhibitory. b) Glutamate is excitatory, while GABA is inhibitory. c) Neither, they are both inhibitory. |
Q17: Which statement best describes the characteristics and complexities of ionotropic and metabotropic receptors in synaptic transmission?
* Ionotropic: Direct action on ion channels; activation of ionotropic receptors causes rapid opening of ion channels, resulting in fast synaptic transmission. * Metabotropic: G-protein mediated action; activation of metabotropic receptors leads to a range of downstream effects, often indirectly controlling ion channels, resulting in slower but longer-lasting effects. **both are describing receptors | A17: a) Metabotropic effects are fast but brief and ionotropic effects are slow but long-lasting because they directly control ion channels. b) Ionotropic effects are fast but brief because they directly control ion channels, whereas metabotropic effects are slow but long-lasting due to their complex signalling mechanism involving G-proteins that indirectly control ion channels. c) Both ionotropic and metabotropic receptors directly control ion channels, leading to fast and brief effects in synaptic transmission. d) Ionotropic receptors are more complex in their signalling mechanism than metabotropic receptors because they can trigger many downstream effects from one receptor. |
Q18: Regarding negative reinforcement: Something is _____ from the environment, that causes the behaviour to increase in frequency, therefore that something must have been _______ | A18: a) added, unpleasant b) removed, pleasant c) removed, unpleasant |
Q19: Which of the following examples correctly illustrates negative reinforcement? a- positive reinforcement, where the addition of warmth (pleasant stimulus) encourages a behaviour. b- removing an unpleasant stimulus (loud alarm), leading to an increase in the behaviour (getting out of bed) due to negative reinforcement. c- negative punishment, where a pleasant stimulus (video game privileges) is removed to decrease an undesirable behaviour (coming home late). | A19: a) Bailey starts wearing earmuffs whenever he goes outside during winter, which adds warmth and makes him feel comfortable. As a result, he starts going outside more often. b) Lee turns off the loud alarm every morning by getting out of bed, which stops the unpleasant noise. As a result, she finds it easier to get out of bed when the alarm rings. c) Alex is late coming home, so his parents take away his video game privileges for the weekend. As a result, he starts coming home on time.
|
Q20: What is the main principle of the incentive-sensitization theory of addiction? *increased sensitivity to the incentive/motivational properties of drug *sensitisation of wanting (mesolimbic/mesocortical) | A20: a) wanting and liking the drug become neutral b) less wanting and more liking of the drug c) more wanting and less liking of the drug |
STUDY TIPS:
→ Build a conceptual framework of the content/materials by structuring your summary notes around key terms and concepts (this can be in any format you find best for you: e.g. table, figures, notebook etc).
→ Specifically, start with key terms/concepts and add examples, theories, and more detailed info. This strategy is known as ‘chunking’ in cognitive psychology and helps manage the limited capacity of working memory, making complex information (such as this unit) easier to remember.
Sequence of an Action Potential
Depolarization (Role of Na+):
What Happens: Voltage-gated sodium channels in the neuron's membrane open, allowing Na+ ions, which are initially at higher concentrations outside the neuron, to rush into the cell, causing the membrane potential to become more positive.
Effect on Membrane Potential: The influx of Na+ ions, which are positively charged, decreases the internal negativity of the neuron (makes the inside less negative). This is called depolarization, and it is what initiates the action potential.
Result: The neuron's interior becomes temporarily more positive, reaching a threshold that triggers the action potential, which is essential for nerve signal transmission.
2. Action Potential:
The depolarization reaches a threshold, triggering an action potential. This is the point at which the neuron rapidly becomes even more positive, leading to the peak of the action potential.
What Happens: The depolarization reaches a threshold, triggering an action potential. Voltage-gated sodium channels continue to open, causing rapid entry of Na+ ions and making the neuron even more positive.
Effect on Membrane Potential: The inside of the neuron becomes significantly more positive compared to the outside, leading to the peak of the action potential.
Result: The action potential is generated, and this electrical impulse travels down the axon to propagate the signal to the next cell in the neural pathway.
Repolarization (Role of K+):
What Happens: After the peak of the action potential, potassium channels open, and K+ ions flow out of the neuron, bringing the membrane potential back toward its resting negative value.
Effect on Membrane Potential: The exit of positively charged K+ ions helps restore the negative charge inside the cell. This movement back toward a negative internal environment is known as repolarization.
Result: The neuron's membrane potential returns toward its resting state. This phase may be followed by a brief period where the inside becomes even more negative than the normal resting potential, a phase known as hyperpolarization, before stabilizing at the resting level.
Refractory Period:
Following repolarization, there is a refractory period where the neuron temporarily becomes hyperpolarized (slightly more negative than the resting potential) and cannot fire another action potential. This phase ensures unidirectional propagation of the action potential along the axon and prevents immediate reactivation.
What Happens: Following repolarization, the neuron enters a refractory period where voltage-gated sodium channels become temporarily inactivated, and potassium channels may remain open, causing hyperpolarization.
Effect on Membrane Potential: During this period, the neuron's membrane potential becomes slightly more negative than the normal resting potential, making it harder for the neuron to fire another action potential.
Result: The refractory period ensures unidirectional propagation of the action potential along the axon and prevents immediate reactivation, thereby allowing the neuron to return to a resting state and get ready for the next signal. This contributes to the controlled timing of neural signalling and prevents overlap of action potentials.
Differences Between Chemical and Electrical Synapses
Chemical Synapses:
● How They Work: Chemical synapses use neurotransmitters to convey messages between neurons. When an action potential reaches the end of a presynaptic neuron, it triggers the release of neurotransmitters into the synaptic cleft. These chemical messengers bind to receptors on the postsynaptic neuron, which then generates a response.
● Examples of Neurotransmitters:
o Dopamine: Involved in reward and pleasure pathways.
o Serotonin: Involved in mood regulation.
o Acetylcholine: Plays a role in muscle activation and memory.
● Hormonal Influence: Some neurotransmitters also function as hormones in the body (e.g., epinephrine, norepinephrine, dopamine).
● Feature: This type of synapse is common in the central and peripheral nervous systems, allowing for complex modulation and control of responses. Chemical synapses are slower but allow for more varied responses.
● Pathways:
o Hormones travel via blood
o NT travel via synapses
Electrical Synapses:
● How They Work: Electrical synapses use gap junctions to allow ions to pass directly from one cell to another, creating an electrical current that continues the action potential in the next neuron.
● Examples:
o Gap Junctions in the Brain: Found in areas requiring synchronized activity, such as in certain parts of the brain that control rhythmic activities like breathing.
● Feature: Electrical synapses allow for rapid communication between cells and are typically faster than chemical synapses. They are often used in pathways where a quick, coordinated response is necessary.
Differences Between Classical and Operant Conditioning
Classical Conditioning:
● How It Works: Classical conditioning involves learning through association between stimuli. An unconditioned stimulus (US) that naturally triggers a response becomes associated with a neutral stimulus (CS), which then becomes capable of triggering a similar response.
● Examples:
o Pavlov's Dogs: Bell (CS) paired with food (US) leads to salivation.
o Fear Response: Loud noise (US) paired with light (CS) leads to startle response.
● NW: the CS begins as a NS, only after being paired with an unconditioned stimulus (US) does the neutral stimulus become conditioned to elicit a conditioned response (CR).
● Feature: This type of learning is automatic and involuntary, involving reflexive behaviours. The response occurs regardless of consequences.
● Pathways:
o US → UR: The unconditioned stimulus naturally triggers the unconditioned response.
o CS + US Pairing: The neutral stimulus is repeatedly paired with the unconditioned stimulus.
o CS → CR: After the association is learned, the conditioned stimulus elicits the conditioned response.
Operant Conditioning:
● How It Works: Operant conditioning involves learning through consequences of voluntary behaviours. Behaviours followed by reinforcement tend to increase, while those followed by punishment tend to decrease.
● Types of Consequences:
o Positive Reinforcement: Adding pleasant stimulus increases behaviour.
o Negative Reinforcement: Removing unpleasant stimulus increases behaviour.
o Positive Punishment: Adding unpleasant stimulus decreases behaviour.
o Negative Punishment: Removing pleasant stimulus decreases behaviour
● Feature: This type of learning is voluntary and involves conscious choices. The behaviour is modified based on its consequences, allowing for more complex behaviour modification.
● Schedules of Reinforcement:
o Fixed Ratio: Reinforcement given after a set number of responses (e.g., every 5 responses).
o Variable Ratio: Reinforcement given after an unpredictable number of responses (e.g., average of 5 responses).
o Fixed Interval: Reinforcement is given for the first response that occurs after a fixed amount of time has elapsed (e.g., every 5 minutes).
o Variable Interval: Reinforcement is given for the first response that occurs after a varying amount of time has elapsed (e.g., an average of 5 minutes).