Infants' Detection of Speech in Noise

1981 ◽  
Vol 24 (2) ◽  
pp. 202-206 ◽  
Author(s):  
Sandra E. Thehub ◽  
Dale Bull ◽  
Bruce A. Schneider

Localization responses to a speech phrase masked by white noise were obtained from infants 6, 12, 18, and 24 months of age and from adults. The masking noise was presented continuously from two loudspeakers located 45° to each side of the infant. During a trial the speech phrase was presented through one of the loudspeakers. A head turn to the signal (correct response) was rewarded by activating an animated toy on top of the speaker. The intensity of the signal was varied over trials (method of constant stimuli) to determine thresholds (defined as the intensity corresponding to 65% correct head tunas) at each of two levels of masking noise, 42 and 60 dBC. Thresholds for the speech signal were comparable across all infant groups for both levels of masking noise. Increasing the masking noise from 42 to 60 dBC resulted in a threshold shift of comparable magnitude for infants and adults. However, adult thresholds were approximately 10–12 dB lower than those of infants at both masking levels.

1976 ◽  
Vol 19 (2) ◽  
pp. 216-224 ◽  
Author(s):  
James T. Yates ◽  
Jerry D. Ramsey ◽  
Jay W. Holland

The purpose of this study was to compare the damage risk of 85 and 90 dBA of white noise for equivalent full-day exposures. The damage risk of the two noise levels was determined by comparing the temporary threshold shift (TTS) of 12 subjects exposed to either 85 or 90 dBA of white noise for equivalent half- and full-day exposures. TTS was determined by comparing the pre- and postexposure binaural audiograms of each subject at 1, 2, 3, 4, 6, and 8 kHz. It was concluded that the potential damage risk, that is, hazardous effect, of 90 dBA is greater than 85 dBA of noise for equivalent full-day exposures. The statistical difference between the overall effects of equivalent exposures to 85 dBA as compared to 90 dBA of noise could not be traced to any one frequency. The damage risk of a full-day exposure to 85 dBA is equivalent to that of a half-day exposure to 90 dBA of noise. Within the limits of this study, TTS t was as effective as TTS 2 for estimating the damage risk of noise exposure.


2016 ◽  
Vol 55 (8) ◽  
pp. 463-471 ◽  
Author(s):  
Harvey Dillon ◽  
Elizabeth Francis Beach ◽  
John Seymour ◽  
Lyndal Carter ◽  
Maryanne Golding

2019 ◽  
Vol 63 (2) ◽  
pp. 381-403 ◽  
Author(s):  
Giovanna Morini ◽  
Rochelle S. Newman

The question of whether bilingualism leads to advantages or disadvantages in linguistic abilities has been debated for many years. It is unclear whether growing up with one versus two languages is related to variations in the ability to process speech in the presence of background noise. We present findings from a word recognition and a word learning task with monolingual and bilingual adults. Bilinguals appear to be less accurate than monolinguals at identifying familiar words in the presence of white noise. However, the bilingual “disadvantage” identified during word recognition is not present when listeners were asked to acquire novel word-object relations that were trained either in noise or in quiet. This work suggests that linguistic experience and the demands associated with the type of task both play a role in the ability for listeners to process speech in noise.


2019 ◽  
Author(s):  
Paul Glad Mihai ◽  
Nadja Tschentscher ◽  
Katharina von Kriegstein

AbstractRecognising speech in background noise is a strenuous daily activity, yet most humans can master it. A mechanistic explanation of how the human brain deals with such sensory uncertainty is the Bayesian Brain Hypothesis. In this view, the brain uses a dynamic generative model to simulate the most likely trajectory of the speech signal. Such simulation account can explain why there is a task-dependent modulation of sensory pathway structures (i.e., the sensory thalami) for recognition tasks that require tracking of fast-varying stimulus properties (i.e., speech) in contrast to relatively constant stimulus properties (e.g., speaker identity) despite the same stimulus input. Here we test the specific hypothesis that this task-dependent modulation for speech recognition increases in parallel with the sensory uncertainty in the speech signal. In accordance with this hypothesis, we show—by using ultra-high-resolution functional magnetic resonance imaging in human participants—that the task-dependent modulation of the left primary sensory thalamus (ventral medial geniculate body, vMGB) for speech is particularly strong when recognizing speech in noisy listening conditions in contrast to situations where the speech signal is clear. Exploratory analyses showed that this finding was specific to the left vMGB; it was not present in the midbrain structure of the auditory pathway (left inferior colliculus, IC). The results imply that speech in noise recognition is supported by modifications at the level of the subcortical sensory pathway providing driving input to the auditory cortex.


2013 ◽  
Vol 380-384 ◽  
pp. 3618-3622
Author(s):  
Kang Liu ◽  
Jian Zheng Cheng ◽  
Li Cheng

There are strong dependencies between wavelet coefficients of speech signal,in this article,based on that,a new corresponding nonlinear threshold function derived in Bayesian framework is proposed to decrease the effect of the ambient noise.Analysis of the data shows the effectiveness of the proposed method that it removes white noise more effectually and gets better edge preservation.


1984 ◽  
Vol 49 (3) ◽  
pp. 303-308 ◽  
Author(s):  
Gail D. Chermak ◽  
Joan E. Dengerink ◽  
Harold A. Dengerink

Cumulative auditory effects of repeated exposure to a 3-min 110 dB SPL white noise were examined in 20 subjects (10 male, 10 female). Statistical analysis revealed greater 8-kHz temporary threshold shift in female ears despite equivalent preexposure 8-kHz thresholds across gender. Cumulative effects were seen in thresholds of octave masking scores, especially in female subjects, possibly reflecting increased susceptibility following the second noise exposure.


1967 ◽  
Vol 10 (4) ◽  
pp. 859-864 ◽  
Author(s):  
Charles Speaks ◽  
Jane L. Karmen

The effect of masking noise on identification of synthetic sentences was studied on seven listeners with normal hearing. Performance-intensity (P-I) functions were defined for speech presentation levels ranging from 20 to 70 dB SPL. Noise exerted no appreciable influence on the steepness of the functions. The shift in threshold of identification varied in direct proportion to the overall noise level for noise intensities exceeding 40 dB SPL. Results are compared with the data of Hawkins and Stevens (1950) for thresholds of intelligibility and detectability and contrasted with previously reported results using a competing speech signal.


1996 ◽  
Vol 8 (2) ◽  
pp. 144-148
Author(s):  
Manabu Ishihara ◽  
◽  
Jun Shirataki ◽  

In this study, a signal was synthesized by removing a speech signal at a certain uniform interval and inserting noise into those signal–absent parts. An auditory experiment was conducted to make clear how humans can hear such synthesized signals. In other words, the relationship between the size of noise and the intensity of signal sound and the relationship between the size of noise and clearness degree were made clear. On the basis of the result of the experiment, in case the size of the white noise inserted is smaller than OdB, a degree of sentence comprehension of over 90 percent is obtained as long as the removed intervals amount to around 60 to 50 percent. In this case, the degree of sentence comprehension is seen to have improved by over 30 percent, in view of the fact that the single syllable comprehension is around 50 to 60 percent. Starting with the region where the removed intervals exceed 50 percent, the degree of sentence comprehension goes down sharply, but this is considered to be due to an effect of the insertion of the white noise. On the basis of the results of this experiment, one of the auditory characteristics to be realized by a digital circuit was made clear.


2018 ◽  
Vol 61 (4) ◽  
pp. 957-972
Author(s):  
Alexandre Hennequin ◽  
Amélie Rochet-Capellan ◽  
Silvain Gerber ◽  
Marion Dohen

Purpose This work evaluates whether seeing the speaker's face could improve the speech intelligibility of adults with Down syndrome (DS). This is not straightforward because DS induces a number of anatomical and motor anomalies affecting the orofacial zone. Method A speech-in-noise perception test was used to evaluate the intelligibility of 16 consonants (Cs) produced in a vowel–consonant–vowel context (Vo = /a/) by 4 speakers with DS and 4 control speakers. Forty-eight naïve participants were asked to identify the stimuli in 3 modalities: auditory (A), visual (V), and auditory–visual (AV). The probability of correct responses was analyzed, as well as AV gain, confusions, and transmitted information as a function of modality and phonetic features. Results The probability of correct response follows the trend AV > A > V, with smaller values for the DS than the control speakers in A and AV but not in V. This trend depended on the C: the V information particularly improved the transmission of place of articulation and to a lesser extent of manner, whereas voicing remained specifically altered in DS. Conclusions The results suggest that the V information is intact in the speech of people with DS and improves the perception of some phonetic features in Cs in a similar way as for control speakers. This result has implications for further studies, rehabilitation protocols, and specific training of caregivers. Supplemental Material https://doi.org/10.23641/asha.6002267


Sign in / Sign up

Export Citation Format

Share Document