Speech production is a process initiated by the speaker formulating a message in their mind and subsequently transmitting it to the listener via speech.
Key aspects of speech:
Conveys meaning
Expresses emotion
Provides information about the speaker (e.g., identity, health, and status)
Anatomy of Speech
Speech production involves several steps:
Selection of spoken words
Formulation of phonetics (sounds)
Articulation by the motor system in the vocal apparatus
Source-Filter Theory:
The lungs provide the energy (airflow).
The vocal folds in the larynx act as the source of sound through vibration.
The oral and nasal cavities function as a filter, resonator, and articulator to shape the sound into recognizable speech.
Important distinction:
Speech production differs from language production, as language can be produced manually through signs as well as spoken.
Speech Motor Control
Illustrative studies (e.g., Real-time MRI of human beatboxing) demonstrate the complex coordination of the tongue, lips, and soft palate.
Neural control involves specific brain regions:
Broca’s area: Traditionally associated with speech planning.
Primary Motor Cortex: Responsible for executing the physical movements of articulation.
Lemma level: Abstract word forms in the mental lexicon containing grammatical information (e.g., gender, tense).
Lexeme level: Word forms regarding morphemes and phonemic segments.
Self-Monitoring: The model includes a “loop” where speakers monitor their own pre-speech (inner voice) to catch errors before they are voiced.
Evidence for model through the “tip-of-the-tongue” (TOT) phenomenon identifying failure in lexeme access after lemma activation.
Spreading-Activation Theory
Highlights that speech planning happens simultaneously across four processing levels:
Semantic: The intended meaning.
Syntactic: Grammatical structure of planned utterance.
Morphological: Morphemes in the sentence.
Phonological: Phonemes within the planned utterance.
Activation flows down from the semantic level, but can also spread “backwards,” explaining why we make mistakes like saying a word that sounds similar to the target.
Prosodic Cues: The "music" of speech, which includes:
Pitch: High or low frequency of voice.
Loudness: Volume used for emphasis.
Duration: The length of time a sound is held.
Listening in Communication
Importance of Common Ground
Shared knowledge, beliefs, and assumptions enhance effective communication.
Listeners' interpretation is often influenced by their own knowledge instead of the speaker's perspective, which may lead to egocentric misunderstandings.
Research Findings on Common Ground
Horton & Keysar (1996) demonstrated that interactive communication reduces error rates in shared tasks (39% errors without interaction vs. 5% with interaction).
Models of Speech Planning (Horton & Keysar, 1996)
Two approaches to planning:
Initial Design Model: Plans are constructed specifically based on shared common ground with the listener from the start.
Monitoring and Adjustment Model: Initial plans focus on the speaker's own perspective (egocentric), followed by corrections based on feedback and knowledge shared with the listener.
Research suggests speakers most commonly use the Monitoring and Adjustment Model because it is less cognitively taxing initially.
Establishing Common Ground (Bard et al., 2007)
Participants demonstrate reliance on:
Shared Responsibility Strategy: Expectation that the listener will signal if they don't understand.
Cognitive Overload: The difficulty of tracking personal knowledge while simultaneously tracking the listener's knowledge.