On the articulation between acoustic and semantic uncertainty in speech perception: Investigating the interaction between sources of information in perceptual classification.

2019 ◽  
Vol 145 (3) ◽  
pp. 1914-1914
Author(s):  
Olivier Crouzet ◽  
Etienne Gaudrain
1990 ◽  
Vol 1 (1) ◽  
pp. 55-63 ◽  
Author(s):  
Dominic W. Massaro ◽  
Michael M. Cohen

The research reported in this paper uses novel stimuli to study how speech perception is influenced by information presented to ear and eye. Auditory and visual sources of information (syllables) were synthesized and presented in isolation or in factorial combination. A five-step continuum between the syllables ibal and idal was synthesized along both auditory and visual dimensions, by varying properties of the syllable at its onset. The onsets of the second and third formants were manipulated in the audible speech. For the visible speech, the shape of the lips and the jaw position at the onset of the syllable were manipulated. Subjects’ identification judgments of the test syllables presented on videotape were influenced by both auditory and visual information. The results were used to test between a fuzzy logical model of speech perception (FLMP) and a categorical model of perception (CMP). These tests indicate that evaluation and integration of the two sources of information makes available continuous as opposed to just categorical information. In addition, the integration of the two sources appears to be nonadditive in that the least ambiguous source has the largest impact on the judgment. The two sources of information appear to be evaluated, integrated, and identified as described by the FLMP-an optimal algorithm for combining information from multiple sources. The research provides a theoretical framework for understanding the improvement in speech perception by hearing-impaired listeners when auditory speech is supplemented with other sources of information.


1989 ◽  
Vol 12 (4) ◽  
pp. 741-755 ◽  
Author(s):  
Dominic W. Massaro

AbstractThis book is about the processing of information in face-to-face communication when a speaker makes both audible and visible information available to a perceiver. Both auditory and visual sources of information are evaluated and integrated to achieve speech perception. The evaluation of the information source provides information about the strength of alternative interpretations, rather than just all-or-none categorical information, as claimed by “categorical perception” theory. Information sources are evaluated independently; the integration process insures that the least ambiguous sources have the most influences on the judgment. Similar processes occur in a variety of other behaviors, ranging from personality judgments and categorization to sentence interpretation and decision making. The experimental results are consistent with a fuzzy logical model of perception, positing three operations in perceptual (primary) recognition: feature evaluation, feature integration, and pattern classification. Continuously valued features are first evaluated, then integrated and matched against prototype descriptions in memory; finally, an identification decision is made on the basis of the relative goodness-of-match of the stimulus information with the relevant prototype descriptions.


2015 ◽  
Vol 58 (3) ◽  
pp. 1093-1102 ◽  
Author(s):  
Brent Spehar ◽  
Stacey Goebel ◽  
Nancy Tye-Murray

Purpose This study compared the use of 2 different types of contextual cues (sentence based and situation based) in 2 different modalities (visual only and auditory only). Method Twenty young adults were tested with the Illustrated Sentence Test (Tye-Murray, Hale, Spehar, Myerson, & Sommers, 2014) and the Speech Perception in Noise Test (Bilger, Nuetzel, Rabinowitz, & Rzeczkowski, 1984; Kalikow, Stevens, & Elliott, 1977) in the 2 modalities. The Illustrated Sentences Test presents sentences with no context and sentences accompanied by picture-based situational context cues. The Speech Perception in Noise Test presents sentences with low sentence-based context and sentences with high sentence-based context. Results Participants benefited from both types of context and received more benefit when testing occurred in the visual-only modality than when it occurred in the auditory-only modality. Participants' use of sentence-based context did not correlate with use of situation-based context. Cue usage did not correlate between the 2 modalities. Conclusions The ability to use contextual cues appears to be dependent on the type of cue and the presentation modality of the target word(s). In a theoretical sense, the results suggest that models of word recognition and sentence processing should incorporate the influence of multiple sources of information and recognize that the 2 types of context have different influences on speech perception. In a clinical sense, the results suggest that aural rehabilitation programs might provide training to optimize use of both kinds of contextual cues.


eLife ◽  
2018 ◽  
Vol 7 ◽  
Author(s):  
Muge Ozker ◽  
Daniel Yoshor ◽  
Michael S Beauchamp

Human faces contain multiple sources of information. During speech perception, visual information from the talker’s mouth is integrated with auditory information from the talker's voice. By directly recording neural responses from small populations of neurons in patients implanted with subdural electrodes, we found enhanced visual cortex responses to speech when auditory speech was absent (rendering visual speech especially relevant). Receptive field mapping demonstrated that this enhancement was specific to regions of the visual cortex with retinotopic representations of the mouth of the talker. Connectivity between frontal cortex and other brain regions was measured with trial-by-trial power correlations. Strong connectivity was observed between frontal cortex and mouth regions of visual cortex; connectivity was weaker between frontal cortex and non-mouth regions of visual cortex or auditory cortex. These results suggest that top-down selection of visual information from the talker’s mouth by frontal cortex plays an important role in audiovisual speech perception.


2020 ◽  
Vol 63 (4) ◽  
pp. 1270-1281
Author(s):  
Leah Fostick ◽  
Riki Taitelbaum-Swead ◽  
Shulamith Kreitler ◽  
Shelly Zokraut ◽  
Miriam Billig

Purpose Difficulty in understanding spoken speech is a common complaint among aging adults, even when hearing impairment is absent. Correlational studies point to a relationship between age, auditory temporal processing (ATP), and speech perception but cannot demonstrate causality unlike training studies. In the current study, we test (a) the causal relationship between a spatial–temporal ATP task (temporal order judgment [TOJ]) and speech perception among aging adults using a training design and (b) whether improvement in aging adult speech perception is accompanied by improved self-efficacy. Method Eighty-two participants aged 60–83 years were randomly assigned to a group receiving (a) ATP training (TOJ) over 14 days, (b) non-ATP training (intensity discrimination) over 14 days, or (c) no training. Results The data showed that TOJ training elicited improvement in all speech perception tests, which was accompanied by increased self-efficacy. Neither improvement in speech perception nor self-efficacy was evident following non-ATP training or no training. Conclusions There was no generalization of the improvement resulting from TOJ training to intensity discrimination or generalization of improvement resulting from intensity discrimination training to speech perception. These findings imply that the effect of TOJ training on speech perception is specific and such improvement is not simply the product of generally improved auditory perception. It provides support for the idea that temporal properties of speech are indeed crucial for speech perception. Clinically, the findings suggest that aging adults can be trained to improve their speech perception, specifically through computer-based auditory training, and this may improve perceived self-efficacy.


2020 ◽  
Vol 29 (2) ◽  
pp. 259-264 ◽  
Author(s):  
Hasan K. Saleh ◽  
Paula Folkeard ◽  
Ewan Macpherson ◽  
Susan Scollie

Purpose The original Connected Speech Test (CST; Cox et al., 1987) is a well-regarded and often utilized speech perception test. The aim of this study was to develop a new version of the CST using a neutral North American accent and to assess the use of this updated CST on participants with normal hearing. Method A female English speaker was recruited to read the original CST passages, which were recorded as the new CST stimuli. A study was designed to assess the newly recorded CST passages' equivalence and conduct normalization. The study included 19 Western University students (11 females and eight males) with normal hearing and with English as a first language. Results Raw scores for the 48 tested passages were converted to rationalized arcsine units, and average passage scores more than 1 rationalized arcsine unit standard deviation from the mean were excluded. The internal reliability of the 32 remaining passages was assessed, and the two-way random effects intraclass correlation was .944. Conclusion The aim of our study was to create new CST stimuli with a more general North American accent in order to minimize accent effects on the speech perception scores. The study resulted in 32 passages of equivalent difficulty for listeners with normal hearing.


2020 ◽  
Vol 63 (7) ◽  
pp. 2245-2254 ◽  
Author(s):  
Jianrong Wang ◽  
Yumeng Zhu ◽  
Yu Chen ◽  
Abdilbar Mamat ◽  
Mei Yu ◽  
...  

Purpose The primary purpose of this study was to explore the audiovisual speech perception strategies.80.23.47 adopted by normal-hearing and deaf people in processing familiar and unfamiliar languages. Our primary hypothesis was that they would adopt different perception strategies due to different sensory experiences at an early age, limitations of the physical device, and the developmental gap of language, and others. Method Thirty normal-hearing adults and 33 prelingually deaf adults participated in the study. They were asked to perform judgment and listening tasks while watching videos of a Uygur–Mandarin bilingual speaker in a familiar language (Standard Chinese) or an unfamiliar language (Modern Uygur) while their eye movements were recorded by eye-tracking technology. Results Task had a slight influence on the distribution of selective attention, whereas subject and language had significant influences. To be specific, the normal-hearing and the d10eaf participants mainly gazed at the speaker's eyes and mouth, respectively, in the experiment; moreover, while the normal-hearing participants had to stare longer at the speaker's mouth when they confronted with the unfamiliar language Modern Uygur, the deaf participant did not change their attention allocation pattern when perceiving the two languages. Conclusions Normal-hearing and deaf adults adopt different audiovisual speech perception strategies: Normal-hearing adults mainly look at the eyes, and deaf adults mainly look at the mouth. Additionally, language and task can also modulate the speech perception strategy.


2020 ◽  
Vol 63 (2) ◽  
pp. 487-498
Author(s):  
Puisan Wong ◽  
Man Wai Cheng

Purpose Theoretical models and substantial research have proposed that general auditory sensitivity is a developmental foundation for speech perception and language acquisition. Nonetheless, controversies exist about the effectiveness of general auditory training in improving speech and language skills. This research investigated the relationships among general auditory sensitivity, phonemic speech perception, and word-level speech perception via the examination of pitch and lexical tone perception in children. Method Forty-eight typically developing 4- to 6-year-old Cantonese-speaking children were tested on the discrimination of the pitch patterns of lexical tones in synthetic stimuli, discrimination of naturally produced lexical tones, and identification of lexical tone in familiar words. Results The findings revealed that accurate lexical tone discrimination and identification did not necessarily entail the accurate discrimination of nonlinguistic stimuli that followed the pitch levels and pitch shapes of lexical tones. Although pitch discrimination and tone discrimination abilities were strongly correlated, accuracy in pitch discrimination was lower than that in tone discrimination, and nonspeech pitch discrimination ability did not precede linguistic tone discrimination in the developmental trajectory. Conclusions Contradicting the theoretical models, the findings of this study suggest that general auditory sensitivity and speech perception may not be causally or hierarchically related. The finding that accuracy in pitch discrimination is lower than that in tone discrimination suggests that comparable nonlinguistic auditory perceptual ability may not be necessary for accurate speech perception and language learning. The results cast doubt on the use of nonlinguistic auditory perceptual training to improve children's speech, language, and literacy abilities.


2013 ◽  
Vol 20 (3) ◽  
pp. 91-106 ◽  
Author(s):  
Rachel Pizarek ◽  
Valeriy Shafiro ◽  
Patricia McCarthy

Computerized auditory training (CAT) is a convenient, low-cost approach to improving communication of individuals with hearing loss or other communicative disorders. A number of CAT programs are being marketed to patients and audiologists. The present literature review is an examination of evidence for the effectiveness of CAT in improving speech perception in adults with hearing impairments. Six current CAT programs, used in 9 published studies, were reviewed. In all 9 studies, some benefit of CAT for speech perception was demonstrated. Although these results are encouraging, the overall quality of available evidence remains low, and many programs currently on the market have not yet been evaluated. Thus, caution is needed when selecting CAT programs for specific patients. It is hoped that future researchers will (a) examine a greater number of CAT programs using more rigorous experimental designs, (b) determine which program features and training regimens are most effective, and (c) indicate which patients may benefit from CAT the most.


Sign in / Sign up

Export Citation Format

Share Document