Perceptual Similarity and distance CGSC433

0.0(0)
studied byStudied by 0 people
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
Card Sorting

1/85

encourage image

There's no tags or description

Looks like no tags are added yet.

Study Analytics
Name
Mastery
Learn
Test
Matching
Spaced

No study sessions yet.

86 Terms

1
New cards

when speech sounds differ from one another in more than 1 acoustic dimension

may be more useful to express confusion rate with perceptual similarity

2
New cards

the perceptual similarity between speech sounds can be

measured

3
New cards

these speech sounds can be put on a

map

4
New cards

sounds closer together are

perceptually more similar

5
New cards

sounds farther apart are

perceptually more distinct

6
New cards

English fricative perception Miller and Nicely 1955

subjects are played [f] and [θ] and asked to identify the sounds as "f" or "th"

<p>subjects are played [f] and [θ] and asked to identify the sounds as "f" or "th"</p>
7
New cards

when investigating speech sounds that differ along multiple (unknown) dimensions

the "confusion scores" give us a way to think about their similarity

8
New cards

when one sound is easily confused with another

we can think of the two sounds as being close together in perceptual space, they are similar

9
New cards

when one sound is not easily confused with another

we can think of the two sounds as being far apart in perceptual space, not similar

10
New cards

these ideas can be made precise with statistical techniques called

Multidimensional scaling (MDS) which is like triangulation, and clustering analyses

11
New cards

essentially multidimensional scaling allows us to

draw a perceptual map in n dimensions where n is any number

12
New cards

we usually use

2 dimensions, so we can see it easily and then we can try to interpret these dimensions

13
New cards

when we make maps, we need to find a way for

all the distances to get along. This is MDS

<p>all the distances to get along. This is MDS</p>
14
New cards

represent perceptual differences between

stimuli as ratio/proportion

15
New cards

estimate

perceptual distance- confusability plus "Shepard's Law"

16
New cards

pin down points in the space with

multidimensional scaling (MDS)

17
New cards

English fricative perception Miller and Nicely data expressed as ratios

knowt flashcard image
18
New cards

calculating proportions from raw data- English fricative perception Miller and Nicely

knowt flashcard image
19
New cards

1st step in finding perceptual distance

determine similarity of these pairs with this equation

<p>determine similarity of these pairs with this equation</p>
20
New cards

S=

similarity

21
New cards

P=

proportion

22
New cards

I and j are

variables representing two sounds under comparison

23
New cards

confusability of [f] with [θ]

knowt flashcard image
24
New cards

confusability of [f] and [θ] continued

knowt flashcard image
25
New cards

confusability of [f] and [θ] answer

knowt flashcard image
26
New cards

S=1 means

exactly the same

27
New cards

S=0 means

totally different

28
New cards

if 2 sounds are exactly the same S=1

perceptual distance should be zero-Shepard's law

29
New cards

if 2 sounds are completely different S=0

perceptual distance should be very large-Shepard's law

30
New cards

perceptual distance is

negative natural log (ln) of similarity: d=-ln(S)

31
New cards

perceptual distance in [f] and [θ]

knowt flashcard image
32
New cards

do all humans perceive speech the same way, regardless of their native language?

no, our own language also affects the way we perceive individual speech sounds

33
New cards

speakers of different languages have

different perceptual maps of the same sounds

34
New cards

perceptual vowel space is similar to

acoustic vowel space

35
New cards

pure auditory differences

matter in all languages

36
New cards

perception of speech can be affected by both

auditory ability and our own language experience

37
New cards

phonetic coherence refers to

our ability to force incoherent signals into a unifying picture--> one example duplex perception

38
New cards

people not only hear the chirp but they also perceive

what is heard from the right ear as [da] or [ga]- the missing piece has been filled in by the ear

39
New cards

people also rely on

visual signals when perceiving speech- example-McGurk effect

40
New cards

McGurk

an error in perception that occurs when we misperceive sounds because the audio and visual parts of the speech are mismatched.

41
New cards

for people who aren't sensitive to the McGurk effect could be for a variety of reasons including

some people are more sensitive to auditory cues than visual cues, bilingualism or learning a second language can change sensory perception

42
New cards

bilinguals experience McGurk effect more than

monolinguals

43
New cards

multisensory speech perception- somatosensation

tactile sensation- in one experiment people heard a syllable starting with either a voiced or voiceless stop, on some trails they also received a puff of air on their hand or neck, speech was presented in masking noise to reduce overall accuracy

44
New cards

the ganong effect

one where your lexical knowledge (the list of words you know) affects your speech perception

45
New cards

more people say they hear dog than gog because

dog is a real word and gog isn't

46
New cards

the real word acts as

a perceptual magnet compared to the non-word

47
New cards

speech perception is more than

acoustic analysis of speech signal, matching analysis with a string of phonemes, putting phoneme string together, identifying word that matches sound sequence, top down processing

48
New cards

lexical knowledge-top down processing helps with

phonetic coherence

49
New cards

sound missing from a speech signal can be

restored by brain and may disappear to be heard

50
New cards

[s] is perceived even when its not there due to

knowledge of word legislation

51
New cards

speech replaced with

3 sine waves that vary in frequency through time

52
New cards

strong evidence that we use knowledge of our own language to

understand speech-top down processing

53
New cards

factors affecting speech perception include

auditory ability, phonetic knowledge, our own language experiences, visual/other sensory cues

54
New cards

despite the great variety of different speakers, different conditions, and coarticulation involved in speech (known as the lack of invariance problem)

listeners perceive vowels and consonants as constant categories

55
New cards

two different approaches

motor theory of speech perception (Liberman et al. 1967, Liberman and Mattingly, 1985) and general auditory approach

56
New cards

motor theory of speech perception basis

relation of phoneme to acoustic is context-dependent

57
New cards

lack of invariance problem

variability problem

58
New cards

general audition and learning cannot account for

uniqueness of human ability to perceive speech, speech-specific linguistic module is needed, speech motor system could provide this

59
New cards

internal synthesizer

mappings between intended articulators/motor commands and acoustic consequences

60
New cards

listeners use the acoustic signal and knowledge of these mappings to

determine intended articulatory gestures of the speaker

61
New cards

speech is understood in terms of

how it is produced

62
New cards

modified--> hypothesized that

articulatory events recovered by human listeners are neuromotor commands to the articulators, rather than actual articulatory movements or gestures

63
New cards

according to the motor theory of speech perception

human ability to perceive speech sounds is special and can't be attributed to general mechanisms of audition and perceptual learning

64
New cards

perception thought to depend on

a specialized decoder or module that is speech-specific, unique to humans, and innately organized as part of human specialization for language

65
New cards

a major benefit of motor theory of speech production

perception and production governed by the same types of units/mechanisms

66
New cards

support for the motor theory of speech perception comes from

McGurk effect-where visual cues of articulation seem to have an effect on the listeners perception of the sound, duplex perception, and speech shadowing was shown to be faster after hearing spoken words

67
New cards

general auditory approaches

acoustic events are the objects of perception

68
New cards

general auditory approaches have

no special-speech module: perception uses the same perceptual systems and learning mechanisms that other animals have

69
New cards

general auditory approach is a term coined by

Diehl and colleagues to describe non-motor alternatives

70
New cards

GAA say context dependence of acoustics is

not a problem: listeners make use of multiple cues and can interpret with acoustic context

71
New cards

support for general auditory approaches

lack of acoustic invariance is not problematic, categorical perception- synthetic sounds can be perceived categorically and non-human animals exhibit categorical perception

72
New cards

drawbacks against general auditory approaches

would be surprising is there were not processes and resources devoted to the perception of speech

73
New cards

research actually supports a special status for speech in auditory perception

evidence suggests that human neonates prefer sounds of speech to non-speech

74
New cards

adults are able to distinguish speech from non-speech based on

visual cues alone

75
New cards

infants 4-6 months can

detect based on visual cues alone, when a speaker changes from one language to another, though all but those in bilingual households lost that ability by 8 months

76
New cards

perceptual judgements by clinicians and even family members of a client play

a major role in understanding diagnosis and management of the disorder

77
New cards

it also matters from the client/patient standpoint

what they can perceive when they have particular communication/hearing disorders

78
New cards

ataxic dysarthria

associated with cerebellar disorders, involves irregular or slow rhythm of speech, with pauses and abrupt explosions of sound and abnormal or excessively equal stress on every syllable

79
New cards

a feature of ataxic dysarthria is

inability to regulate long-short-long-short patterning of syllable durations in multi-syllabic utterance

80
New cards

normal listeners will have trouble

identifying word onset locations in speech of someone with ataxic dysarthria

81
New cards

SLPs understanding of

link between neural disease, speech timing problems, and lexical access as an important part of speech perception is crucial which allows more effective intervention on speech timing issue to benefit both speaker and listener

82
New cards

standardized articulation tests evaluate

goodness of consonant production in word-initial, medial, and final positions such as /s/ in sock, messy, and bus

83
New cards

listeners rely mostly on

word-initial consonants for word identification in continuous speech

84
New cards

audiometric testing

identification of frequencies of hearing loss

85
New cards

significant hearing loss in higher frequencies 4,000-8,000 Hz compared with

lower frequencies 250-2,000 Hz will likely affect processing of sounds like /s/, /ʃ/, and /t/

86
New cards

knowledge of speech perception is important for

matching amplification characteristics of hearing aids to hearing loss patterns across frequency spectrum