SLHS 2010 Final Exam 2

0.0(0)
studied byStudied by 0 people
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
Card Sorting

1/54

encourage image

There's no tags or description

Looks like no tags are added yet.

Study Analytics
Name
Mastery
Learn
Test
Matching
Spaced

No study sessions yet.

55 Terms

1
New cards

What is the difference between lung volume and lung capacity?

Volumes—a way of partitioning off what each part of the lungs can hold

Discrete measures

Capacities—refer to the combinations of volumes that reflect physiological limits

Functional combinations of volumes

2
New cards

What are the differences between speech and life breathing?

Life breathing—passive process

Nose

40% of breathing cycle is inhalation, 60% is exhalation

Depends on

Body's metabolism

Body temperature

Fluid and acid-base balance

An unconscious automatic process

Rhythmic control center of medulla oblongata

Gravity pulls down rib cage

Inhalation and exhalation 1pprox... 2 seconds

Diaphragm and external intercostals relax

Elasticity returns lung tissues to normal

Torque of cartilaginous portions of ribs

Speech breathing—active process

Mouth—more efficient—less distance to lungs

10% of cycle used in inhalation, while 90% of cycle for exhalation

Depends on

Controlled release of inhaled air

Automatic nerve signals for life breathing are incorporated with voluntary nerve signals for speech breathing

You speak when you exhale

Needs more air

A conscious, controlled process

Inhalation time shorter but exhalation much longer

Vocal folds need to vibrate and produce the sawtooth wave

Air being pushed out from lungs is a more controlled air stream

Above REL, muscle control keeps air from escaping too fast

Below REL, expiration muscles contract to continue decreasing volume of thorax and lungs

3
New cards

What occurs during inhalation? What occurs during exhalation? How does this relate to Boyle's law?

Inhalation

Contract diaphragm

Contract external intercostals

Together, these two contractions raise the rib cage and expand the thorax

Given pleural linkage, when the thorax expands, the lungs also expand, drawing in air (Boyle's Law)

Boyle's Law

As the volume of a container increases, air pressure within the container decreases

Exhalation

Diaphragm and external intercostals relax

Internal intercostals contract

Thorax (and lungs) decrease in volume, and air is forced out

4
New cards

What is the one laryngeal bone? Where does it sit in the larynx?

The Hyoid Bone: A small /u-shaped bone at the top of the larynx

Attachment for tongue

5
New cards

How many layers do vocal folds have?

5

6
New cards

What is the myoelastic aerodynamic theory of phonation?

Voice production is a combination of:

Muscle (myo-)

Tissue elasticity (-elastic)

Air pressure (aero-)

Flow (-dynamic)

7
New cards

If given a fundamental frequency at a particular amplitude, you should be able to calculate the frequencies and amplitudes of all the harmonics.

If the fundamental is 100 Hz. At 70 dB SPL, then what are the frequencies and amplitudes of the harmonics?

Freqharm=freqfund x harmonic number

Ampharm=ampfund + [20 x log (1/harmonic number)]

Harmonic #2 will have double the frequency and half the amplitude.

2x100=200 Hz

Ampharm=70 + [20 x log(1/2)]

Ampharm=70+ [20 x-3]

Ampharm=70+ [-6]

Ampharm=64 dB

8
New cards

How do you find the fundamental frequency of a complex harmonic sound?

If the fundamental frequency is 50Hz, then the second would be 100Hz, the third would be 150 Hz, and so on.

9
New cards

How does a speaker change vocal pitch?

We can alter muscular activity to change fundamental frequency. We have conscious control over pitch

Fundamental frequency variability signals sentence type, mood, topicalization, grammatical function, etc.

Lengthening and/or tensing will increase fundamental frequency

Lengthening will decrease mass per unit length

Tensing will also increase rate of VF vibration

Relaxing will decrease fundamental frequency

Relaxing will increase mass per unit length

10
New cards

How does pitch differ between speakers? What are the general trends related to pitch changes due to age and sex?

Infants and toddlers= 400 Hz, shrieks= 1000 Hz (length 5-7 mm)

Older children (4-10 yrs): vocal folds lengthen and thicken (length 8 mm), causing decrease in fundamental frequency (no real sex differences)

Puberty: male larynges become longer and thicker, causing decrease in fundamental frequency. Female larynges undergo similar development, but not as extreme. Learning to adjust/control pitch (length: boys 15-25 mm; girls 12-17 mm)

Adults: Men=120 Hz (~29 mm), Women= 220 Hz (~21 mm)

In later life, male vocal folds become thinner, while female vocal folds become thicker, reducing sex differences (~145 vs. ~175 Hz)

11
New cards

What is the Source Filter Theory? What is the source for voiced sounds? For unvoiced sounds? What is the filter?

Explains speech and how we produce it based on the math we studied

The source- Vocal folds vibrating produces a periodic source (remember a sawtooth wave?)

The source is the output of the larynx

The Filter- The tube that the source sound goes through (remember a closed tube?)

The filter is the resonances of the vocal tract

To produce speech, we need some energy to create vibration

A source (one of the requirements for sound), and something to change it form the buzzy sawtooth to speech (e.g., a filter)

12
New cards

What are the different ways you can measure voice?

Vocal registers (modes of phonation)

The range of voice pitches range from <60 Hz to >1568 Hz

Pulse—low pitch

Modal—normal pitch

Falsetto—high pitch

Vocal Quality

Very broad term

Subjective

E.g., pleasant, strident, rough, raspy, shrill, clear, unpleasant, harsh, hoarse, tinny, strained...

Hyperadduction heard as tense

Hypoadduction heard as noisy or breathy

No standard frame of reference for these

Abnormal Voice Qualities

Vocal quality Is determined by the presence of noise

Breathiness—aspirated vocal tone (strong bursts of air)

Roughness—raspy vocal tone, low pitch

Hoarseness-a combination of breathy and rough qualities

Measuring Voice

Videostroboscopy

A scope that shows a video of vocal fold vibration

Can directly examine appearance, shape, size of vocal folds; look for nodules or lesions

Must be out of phase with voice to see a complete cycle of voicing

Imaging

Computed Tomography (CT)

Magnetic Resonance Imaging (MRI)

Can examine soft tissue

Electroglottography (EGG, laryngography)

A noninvasive technique that measures how tightly the vocal folds touch one another

The more the tissues are in contact, the higher the peaks

Valleys represent times when the vocal folds are separated

13
New cards

What are formants?

The peaks in spectral energy are called formants

Vocal source resonates at the preferred frequencies of the vocal tract

Infinite number of formants possible

5 in clear speech

The first 3 are the most important

Formants are made up of several harmonics that are amplified from resonant filtering

14
New cards

What are active and passive articulators?

Active Articulators-those that do most of the moving during a speech gesture

Lips

Tongue (tip/blade/body/root)

Velum (controls whether air resonates in nasal cavity)

glottis

Passive Articulators- those that do little or no moving during a speech gesture

teeth

Alveolar ridge

Palate

Velum (as a place of articulation)

Uvula

Pharyngeal wall

15
New cards

How does a single speaker produce the different vowels? What movement influences F1? What movement influences F2?

Across speakers, the same vowel has different formant frequencies due to differences in speaker vocal tracts

16
New cards

How do acoustic measurements of vowels relate to articulatory description?

Articulatory Description: How we create the sound by using our articulators

Acoustic Description: What it sounds like. What are the frequencies and amplitudes in the speech sound?

We can describe the vowels in terms of their articulatory descriptions

E.g., /i/ is a high front vowel

These articulatory descriptions perfectly predict the acoustic descriptions

/i/ is a high vowel, so it has a low-frequency first formant

/i/ is also a front vowel, so it has a high-frequency sound formant

17
New cards

What determines differences in formant frequencies between different speakers?

Formant 1 depends on tongue height

The volume of the pharyngeal cavity, and

How tightly the oral cavity is constricted

Formant 2 depends on tongue advancement

Length of the oral cavity

18
New cards

What is articulation? What is the purpose? How is that purpose accomplished?

Modifying the output of the respiratory and phonatory systems to create individual speech sounds

Manipulate resonance by altering the shape of the articulatory cavities

Purpose is to create individual speech sounds

19
New cards

What is the distinction between vowel production and consonant production?

If air from the glottis is allowed to pass through the articulators unimpeded, it is considered a vowel

Motion of articulators changes shape of vocal tract, influencing which frequencies are enhanced/reduced

If air is obstructed, partially or totally, in one or more places, it is considered a consonant

Direct—motion of articulators DIRECTLY influences flow of air stream (constricts/blocks)

20
New cards

What are manner, place, and voicing?

Articulatory Descriptors

Manner of articulation

Place of articulation

Voicing

Manner of Articulation: 11 different manners, English uses 5

Place of Articulation: 17 different places, English uses 8

Voicing

Voicing refers to whether the vocal folds are activated

[-voice] glottis is open, vocal folds are not vibrating

[+voice] vocal folds are vibrating

21
New cards

What are stop consonants? What are nasals? What are fricatives? What are affricates? What are approximants? What are glides?

Stop- airflow is blocked and released as a burst

Nasal- Stop with velum lowered so air passes through the nasal cavity

Fricative- Air forced through narrow channel causing turbulent airflow

Approximant-Articulators approach, but do not touch (approximate one another)

Affricate-produced with a stop-like closure with a fricative release

Glide (or semivowels) - gradual articulatory motions, vowel-like nature

22
New cards

How do tongue tip position and lip position play into vowel production?

High vs. low tongue position affects F1

Front vs. back tongue position affects F2

Tongue tip position affects F3

Lip position also matter somewhat

Lip position affects the overall resonance

Spreading

Rounding

Neutral

23
New cards

What is assimilation?

Phonemes "change" to be more compatible with surrounding phonemes

"in-" prefix in "impossible"

"in-" prefix in "intolerable"

"in-" prefix in "incorrigible"

Nasal /n/ assimilates to same place of articulation as the following phoneme

English past tense "-ed"

"walk" + "-ed" /wɔkt/

"jog" + "-ed" /dʒɔgd/

English plural "-s"

"cat" + "-s" /kæts/

"dog" + "-s" /dɔgz/

Not just in affixes-happens across word boundaries

"ten names" /tɛn:eɪmz/

"ten cards" /tɛŋkaɹdz/

"hard day" /haɹd:eɪ/

"hard book" "harb book" /haɹb:ʊk/

24
New cards

What is coarticulation? How does it affect speech production? How does it affect speech perception?

In conversational speech, we're producing 150-200 words a minute, or about 10 phonemes each second

Because we produce phonemes so rapidly, adjacent phonemes affect one another

They overlap in time and in frequency

The way that any phoneme is produced depends on the phonemes that come before it, and the phonemes that come after it

The positioning of the articulators at any point in time represent a compromise between where the articulators just were and where they're heading

Mushing together sounds

The articulators move toward the configuration appropriate for the upcoming phoneme, even before that sound begins

Lip rounding in "screws" /skɹuz/ vs. "script" /skɹɪpt/

Spread /s/ in "see" vs. rounded /s/ in "sue"

If you try this, you'll see that your lips are in the configuration appropriate for the upcoming sound

Coarticulation has two consequences

Lack of invariance

Because phoneme production varies in terms of surrounding context, there Is no invariant acoustic "fingerprint" for any phoneme

Lack of segmentability

The phonemes are so influenced by what comes before and after that it's impossible to separate a phoneme from surrounding ones

For speech production, coarticulation provides an efficient way to transmit a large amount of information in a relatively short time

For speech perception, it's a complication that has to be disentangled

As speaking rate increases, the phonemes get shorter and overlap more

More likely to "undershoot" articulatory targets, meaning formant frequencies don't hit targets

Some sounds may be completely deleted (elision)

Coarticulation vs. Assimilation

Coarticulation is greater at faster rates and in less careful speech

Assimilation is not dependent on rate or speaking style

25
New cards

What is the lack of segmentability?

The phonemes are so influenced by what comes before and after that it's impossible to separate a phoneme from surrounding ones.

26
New cards

What are the components of prosody?

Prosody is an interaction among intonation, stress and rhythm.

Intonation

Refers to the global pattern of pitch changes that occurs during an utterance

Typical statement—pitch starts high and falls

Typical question—pitch starts low and rises at the end

Can also signal emotion

Stress

Whereas intonation refers to pitch changes that occur over the utterance, stress refers to pitch changes within individual words.

In English, stress is conveyed via pitch, duration, and amplitude

Rhythm

Refers to the relative timing of syllables in the speech stream

In English, stressed syllables tend to have longer duration

Also, words at the end of clauses and sentences tend to have longer duration

27
New cards

What is the difference between conduction and transduction?

Conduction: Outer and middle ears receive acoustic energy from the environment and transmit this sound to the inner ear

Transduction/encoding: Inner ear transduces acoustic energy to electrical energy (neural impulses)

Change energy one form to another

28
New cards

What are the functions of the outer ear? What is the landmark between the outer and middle ear?

Functions of the Outer Ear

Conduction

Outer and middle ears receive acoustic energy from the environment and transmit this sound to the inner ear

Resonance

Our ear canal acts as a closed tube...meaning it acts as a quarter wave resonator

The smaller the ear canal, the greater resonance of higher frequencies

Infants and small children have short ear canals, therefore higher frequencies resonate

This may be why younger children are more bothered by loud high frequency sounds

This may also be why Motherese* is often used

Protection

S-shaped curve of the external auditory canal helps to keep foreign objects out of the ear

Ear wax (cerumen) collects dust and other particles that enter the ear

Localization

The ability to locate where the sound source is located (where the sound is coming from)

When we hear a sound, the two ears get slightly different acoustic versions

The brain compares information from both ears in order to determine where a sound is coming from

Outer Ear Functions

Collects/funnels sound

Protection

Resonance (amplifies certain frequencies)

Localization

ITDs

ILDs

Tympanic membrane—the transition between the outer and middle ear systems

Thin, taut piece of tissue at the end of the ear canal

Vibrates in response to sound waves that are funneled into the ear canal

Those vibrations are then transferred to the middle ear.

29
New cards

Be able to calculate resonant frequencies of the vocal tract and ear canal.

If the average adult ear canal is 2.5 cm long, what is the first resonant frequency (fundamental frequency) that resonates in this system?

344 m/s/4 * 0.025 =3440 Hz

30
New cards

What are the structures in the middle ear? What are the functions of each of these structures?

Middle Ear Functions

Mechanically links outer ear to inner ear

Amplifies sound (impedance matching)

Protects against loud sound (acoustic reflex)

Normalizes pressure

The middle ear is an air-filled space

Contains the ossicular chain

Malleus

Incus

Stapes

Also contains the eustachian tube

Functions

Impedance matching mechanisms

Protects against loud sounds (via the acoustic reflex response)

Pressure normalization

31
New cards

What is the function of the Eustachian tube?

Connects the middle ear to the nasopharynx

Keeps air pressure between middle and outer ears relatively equal

Normally closed, opens to allow air in or out of the middle ear cavity

32
New cards

What is impedance matching? Why is it necessary in the auditory system?

Impedance

Opposition to the flow of energy (e.g. resistance)

Different mediums have different impedances. Air has a lower impedance than fluid.

Energy always wants to move from high impedance to low impedance

The problem arises with conducting sound from the outer ear (air-filled low impedance) to the inner ear (fluid-filled, high impedance)

Without an "impedance matching mechanism," 99.9% of the acoustic energy from the outer ear would be lost by the time it reaches the inner ear.

There are two chambers -middle ear and

The middle ear helps to overcome the impedance mismatch

Different impedances

The Middle Ear over comes impedance mismatch by:

Surface area difference between the tympanic membrane and the oval window of the cochlea

The same amount of force is being concentrated from the tympanic membrane with a large area to the oval window with a small area, so there pressure increases by 17X at the level of the oval window! This means a boost in dB SPL by the time the signal reaches the inner ear.

Lever action of the ossicles

The malleus is longer than the incus, so the malleus moves a greater distance and the incus moves with greater force

Conical shape and flexibility by the tympanic membrane

Helps increase the pressure level by the time sound gets transmitted to the malleus

33
New cards

The middle ear both amplifies sound and attenuates loud sounds? What causes amplification? What causes attenuation?

Amplification: the surface area between the Tympanic Membrane and oval window of cochlea are different. There is a lever action of the ossicles; the malleus is bigger than the incus so it moves with greater force. The conical shape and flexibility of the Tympanic Membrane also aids in amplification.

Attenuation: Acoustic Reflex - the muscles connected to the ossicles contract. There is a decrease in the amount of vibration transmitted across the chain

34
New cards

Attenuation: Acoustic Reflex - the muscles connected to the ossicles contract

There is a decrease in the amount of vibration transmitted across the chain

35
New cards

What are Interaural Timing Differences and Interaural Level Differences? What are they used for?

Inter-aural Level Difference (ILD) Cues:

The brain compares amplitudes of a sound between the two ears

Primarily a result of the head shadow effect

Inter-aural Timing Difference (ITD) Cues:

The brain compares phase difference of a sound between the two ears

Determined by the distance between the two ears

36
New cards

What are the three chambers of the cochlea?

-Scala vestibuli

-Scala media

-Scala tympani

37
New cards

What is the actual organ of hearing? Where in the cochlea is it located? What structure does it rest upon? What are the sensory receptor cells called?

The Organ of Corti

Situated on top of the basilar membrane

The organ of hearing

Contains 4 rows of hair cells:

3 rows of outer hair cells

1 row of inner hair cells

38
New cards

What is the pattern of movement within the cochlea called? Be able to give a brief description of that pattern of movement and describe the acoustic correlates of basilar membrane movement.

Movement of the cochlear fluids causes displacement of the basilar membrane. This displacement occurs in the form of a traveling wave.

Traveling wave

Basilar Membrane vibrates in response to sound

Where the vibration occurs depends on the frequency of the sound

Movement of traveling wave along basilar membrane provides information about

Amplitude

Frequency

Amplitude of sound: Height of the traveling wave. Sound of higher intensity will induce a larger traveling wave

Frequency of sound: Where the traveling wave peaks. Low frequency sound will peak at the apex. High frequency will peak at the base

When the basilar membrane gets displaced (traveling wave occurs), the movement causes tiny cells called stereocilia to shear back and forth along the tectorial membrane

39
New cards

What is tonotopic organization? What happens near the base? What happens near the apex?

The cochlea is organized by frequency

Each point on the basilar membrane is tuned to a particular frequency

The base is narrow and tight, tuned to high frequencies

The apex is wide and loose, tuned to low frequencies

The basilar membrane in the cochlea is tonotopically organized by frequency.

However, not all frequencies are equally represented.

40
New cards

What is the difference between inner hair cells and outer hair cells? What are their functions?

Inner hair cells transduce mechanical energy from vibration into neural energy (an electrical signal)

Outer hair cells help amplify the sound (boost the traveling wave)

41
New cards

Be able to describe the sequence of events that occur during cochlear transduction

-Vibration of stapes pushes on oval window and displaces cochlear fluids in the scala vestibuli and scala tympani

-This movement causes basilar membrane to vibrate and traveling wave to occur

-Movement of basilar membrane causes stereocilia in the hair cells in the organ of corti to shear against tectorial membrane

-This shearing of hair cells causes the hair cells to depolarize

-When hair cells become depolarized, neurotransmitters get released triggering an action potential to auditory nerve

42
New cards

Where does the auditory nerve begin?

Auditory nerve fibers originate at the hair cells of the cochlea

43
New cards

What are the structures in the Central Auditory System (CAS)?

-Auditory nerve

-Brainstem

-Midbrain

-Auditory Cortex

CANS pathway?????

Auditory nerve

Cochlear nucleus

Superior Olivary Complex

Nucleus of the Lateral Lemniscus

Inferior Colloculus

Medial Geniculate Body of the Thalamus

Auditory Cortex

44
New cards

What are the structures in the auditory brainstem and midbrain?

-Cochlear nuclei

-Superior olivary complex

-Lateral lemnscus

-Inferior colliculis

-Medial geniculate body

45
New cards

What is the difference between afferent and efferent fibers? What information is transmitted by both?

-Afferent: send sensory signals away from the cochlea, to the brain

periphery sending information to central nervous system

-Efferent: send information away from the brain, to the cochlea

Exit:

What is the periphery vs what is central

What parts of ear are in periphery

Cochlea to brain

Just like periphery to central

46
New cards

What lobe in the brain is the auditory cortex located in? What is the pattern of organization in the auditory cortex called?

-The auditory cortex lies in the temporal lobe of the brain

-The pattern of organization is called tonotopic organization

47
New cards

What are the three aspects of a sound that are conducted to cortex? How are those three aspects transmitted and encoded by the brain?

-Frequency:

-Tonotopic organization is preserved throughout the entire auditory system, from cochlea to auditory cortex

-The brain knows which frequencies are in the sound based on where the auditory nerve fibers fire (place code)

-Intensity (Amplitude):

Intensity is encoded by how many neurons fire (small group? Large group?) and the rate at which neurons fire (fast? Slow?)

Higher intensity sounds will make large groups of neurons fire at a very fast rate

Low intensity sounds will make small groups of neurons fire at a very slow rate

-Timing:

Conveyed by when neurons fire

Timing is very important for separating detailed information about speech. (e.g., voice onset time, or when the vocal folds vibrate during speech)

48
New cards

How well does neural response to speech match the acoustics of speech? How is loudness related to frequency (e.g., phon contours graph/equal loudness curves)?

What the ear actually gets is an acoustic signal that is very bottom-heavy

A lot of spectral information at lower frequencies

Little spectral information at higher frequencies

Most of the time, we perceive the world as a unified bundle of sensations from multiple sensory modalities (vision, audition, touch). In other words, our perception is multimodal.

Speech is inherently multi-modal.

Human perception of amplitude depends on frequency

Our auditory system is more sensitive to sounds between 1000-5000 Hz

The range of frequencies humans can perceive is limited by the basilar membrane

Apex responds to low frequency sounds, about 20 Hz

Base responds to high frequency sounds, about 20,000 Hz

Frequency range is 20-20,000 Hz

Equal loudness curves depict the physical amplitude (dB SPL) required to make different frequencies the same loudness as a 1000 Hz tone.

49
New cards

How do we use bottom-up processing to perceive speech? How do we use top-down processing?

Bottom-up

cues: The actual sensory input we are receiving (vision, auditory)

when a speaker produces an utterance, s/he is actually conveying two types of information.

Linguistic properties

Language

Carry the speaker's intended message

Indexical properties

Acoustics

Carry information about the speaker's voice

Top-down cues: prior knowledge, context

50
New cards

What is Talker Normalization? What are the two sources of information in the speech stream? Why would we need to focus on one more than the other?

Talker Normalization

The way that any phoneme is produced depends on:

Phonetic variables (i.e., linguistic properties)

Speaker variables (i.e., indexical properties)

We need a way to link the varying realizations of different speech sounds to the underlying units (phonemes)

Through talker normalization

Strips away the indexical properties to get at the linguistic properties

Normalization of Vowels

Vowels vary in terms of formants

We may identify different vowels based on formant frequencies

BUT... remember that formants are determined by vocal tract

Men—Women—Children

Larger people have longer vocal tracts, and consequently, lower formants

The listener hears a single acoustic waveform that contains two sources of information

The absolute values of formant frequencies reveal information about speaker (speaker/indexical variables)

The relative values of formant frequencies reveal information about vowel identity (phonetic/linguistic variables)

Relative values?

Proposed transformations via ratios or logarithms

Pro—We can reduce/eliminate differences due to talker, age, gender

Con—We still get variability

Also, different vowels can have the same ratio

This is a problem...

Maybe the goal is not to strip away talker information, but rather to use it to guide perception

Perception of consonants:

In spite of difficulties, vowel perception is easier than consonant perception

Vowels tend to be voiced, have higher amplitude, clear formants, and longer durations

BUT... consonants:

Have a shorter duration

Are not always voiced

Don't always have clear formants

51
New cards

What is Categorical Perception? What are common tasks that show categorical perception?

Many phonemes are perceived categorically

Stops—voicing [ba]-[pa]

Stops—Place of Articulation [ba]-[da]-[ga]

Glides—F3 Transitions [ra]-[la]

Consonant Clusters—Silence "say"-"stay"

Manner—Amplitude Rise Time [∫a]-[t∫a]

Manner—Oral vs. Nasal [ba]-[ma]

But some are NOT perceived categorically

Fricative Place of Articulation [sa]-[∫a]

Stop vs. Glide Manner [ba]-[wa]

Vowels [i]-[I]

The phenomenon of categorical perception suggests that we perceive in terms of the phonetic categories

Which is learned during development

We can easily tell the difference between two phonemes from different categories, but we ignore the difference between two different versions of the same phoneme

CP of nonspeech sounds

e.g., musical intervals, not tones

CP of visual stimuli

faces, facial expressions

CP of speech by nonhuman animals

Chinchillas (adorbs!!)

52
New cards

Is Categorical Perception the same in all languages? The studies that compared CP in English/Spanish: What did they find? What do those results say about CP?

No, it is no the same in all languages. It depends on your language.

Spanish vs. English

Spanish speakers start voicing sooner than English speakers

This affects how they perceive the same tokens

CP cannot account for all aspects of speech perception

Relying only on bottom-up information is not the whole story

Speech is heard in the presence of information (i.e., other sounds, visual cues, language context, etc.)

We use "other" cues to process speech, too

So, what cues do we use???

Context, top-down processing, speech segmentation, multimodal integration, bootstrapping, statistics/phonotactics

53
New cards

What are the three basic claims of the Motor Theory? What does Motor Theory say about perception of speech vs. perception of other auditory sounds?

Perception is based on production

Even though there are no invariant acoustic features, there are invariant articulatory features that listeners can use to perceive speech.

Perception is species-specific

Since humans are the only species to produce speech, they will be the only ones to perceive it.

Perception is innate

Infants show evidence of perceiving speech well before they can produce it. They do not have to "learn" to perceive the sounds of speech.

We have two distinct sound processors

Speech-phonetic mode

Nonspeech sounds—general auditory mode

54
New cards

What are the three basic claims of the Auditory Theories? What do these theories say about perception of speech vs. perception of other auditory sounds?

1. Perception Is not based on production

Speech perception is like the perception of any auditory sound—General Mechanisms

2. Perception is not species-specific

Any species with a comparable auditory system should perceive speech like humans do

3. Perception may be innate

Because the auditory system is pretty much fully developed at birth, speech perception abilities may be innate

55
New cards

resonant frequency

speed of sound/ 4x the length