Measurement of Visual Speech Comprehension

1970 ◽  
Vol 13 (4) ◽  
pp. 856-860 ◽  
Author(s):  
Henning Birk Nielsen

This study was designed to standardize the measurement of visual speech comprehension in order to facilitate the evaluation of a patient’s communication handicap, his basic capacity for lipreading and the benefit received through lipreading training as a clinical tool. To this end, a silent color film of four minutes' duration was used to obtain a standardized measurement of the lipreading ability of 1108 hearing-impaired subjects. The film depicts an everyday situation in which two persons, while having coffee together, speak nine sentences of varying difficulty. Item analysis reveals good consistency, and the score obtained by the individual patient corresponds well with the clinical evaluation of his lipreading ability.

2012 ◽  
Vol 4 (1) ◽  
pp. 17-24

ABSTRACT The goals of any test that evaluates balance function are manifold; primarily to determine the presence of a true balance disorder and to topographically locate it. Secondarily, to determine the extent of residual functional abilities of the patient as regards the ‘deficit’ caused by the disorder, and evaluate the possibility of recovery. Thirdly, and most importantly, to determine whether the individual is likely to benefit from some therapeutic modality, whether single or combined, based on the results of the tests. Testing for vertiginous patients involves eliciting a detailed history, followed by a clinical evaluation. Investigations include the audiological, radiological and computerized tests. Computerized vestibular testing includes various computerized modules, such as: Electronystagmography (ENG), videonystagmography (VNG), rotational testing, computerized dynamic posturography (sensory organizational test—SOT; motor control testing—MCT) and vestibular evoked myogenic potential (VEMP) testing. VNG is thus only one of the computerized tests and should always be interpreted in conjunction with the others mentioned above. This, however, does not mean that VNG should be performed in all patients complaining of vertigo. It is thus, prudent to understand the indications and possible information that may be obtained from a VNG evaluation and its application to clinical science. Analysis is often carried out by a technician and the results presented to the surgeon or physician. The analysis is often carried out automatically, or at least semi-automatically by a computer, where the automatic artefact rejection is usually poor, or in the least, suboptimal. These potentially ‘weak links’ in the chain can often mislead the clinician, and lead to errant diagnosis, such as ‘central vestibular disorder’, either due to an artifactual recording, or poor interpretation! It is thus, extremely important, that the ENG/VNG results are correlated with the clinical evaluation of the same tests, which may be done with the naked eye or in the least, using Frenzel's glasses. How to cite this article Hathiram BT, Khattar VS. Videonystagmography. Int J Otorhinolaryngol Clin 2012;4(1): 17-24.


2018 ◽  
Vol 116 (2) ◽  
pp. 660-669 ◽  
Author(s):  
Mohsen Alavash ◽  
Sarah Tune ◽  
Jonas Obleser

Speech comprehension in noisy, multitalker situations poses a challenge. Successful behavioral adaptation to a listening challenge often requires stronger engagement of auditory spatial attention and context-dependent semantic predictions. Human listeners differ substantially in the degree to which they adapt behaviorally and can listen successfully under such circumstances. How cortical networks embody this adaptation, particularly at the individual level, is currently unknown. We here explain this adaptation from reconfiguration of brain networks for a challenging listening task (i.e., a linguistic variant of the Posner paradigm with concurrent speech) in an age-varying sample of n = 49 healthy adults undergoing resting-state and task fMRI. We here provide evidence for the hypothesis that more successful listeners exhibit stronger task-specific reconfiguration (hence, better adaptation) of brain networks. From rest to task, brain networks become reconfigured toward more localized cortical processing characterized by higher topological segregation. This reconfiguration is dominated by the functional division of an auditory and a cingulo-opercular module and the emergence of a conjoined auditory and ventral attention module along bilateral middle and posterior temporal cortices. Supporting our hypothesis, the degree to which modularity of this frontotemporal auditory control network is increased relative to resting state predicts individuals’ listening success in states of divided and selective attention. Our findings elucidate how fine-tuned cortical communication dynamics shape selection and comprehension of speech. Our results highlight modularity of the auditory control network as a key organizational principle in cortical implementation of auditory spatial attention in challenging listening situations.


2011 ◽  
Vol 4 (6) ◽  
pp. 1147-1159 ◽  
Author(s):  
A. Richter ◽  
M. Begoin ◽  
A. Hilboll ◽  
J. P. Burrows

Abstract. Satellite observations of nitrogen dioxide (NO2) provide valuable information on both stratospheric and tropospheric composition. Nadir measurements from GOME, SCIAMACHY, OMI, and GOME-2 have been used in many studies on tropospheric NO2 burdens, the importance of different NOx emissions sources and their change over time. The observations made by the three GOME-2 instruments will extend the existing data set by more than a decade, and a high quality of the data as well as their good consistency with existing time series is of particular importance. In this paper, an improved GOME-2 NO2 retrieval is described which reduces the scatter of the individual NO2 columns globally but in particular in the region of the Southern Atlantic Anomaly. This is achieved by using a larger fitting window including more spectral points, and by applying a two step spike removal algorithm in the fit. The new GOME-2 data set is shown to have good consistency with SCIAMACHY NO2 columns. Remaining small differences are shown to be linked to changes in the daily solar irradiance measurements used in both GOME-2 and SCIAMACHY retrievals. In the large retrieval window, a not previously identified spectral signature was found which is linked to deserts and other regions with bare soil. Inclusion of this empirically derived pseudo cross-section significantly improves the retrievals and potentially provides information on surface properties and desert aerosols. Using the new GOME-2 NO2 data set, a long-term average of tropospheric columns was computed and high-pass filtered. The resulting map shows evidence for pollution from several additional shipping lanes, not previously identified in satellite observations. This illustrates the excellent signal to noise ratio achievable with the improved GOME-2 retrievals.


Author(s):  
Gabriella Bekő ◽  
Henriett Butz ◽  
Klára Berta ◽  
András Tislér ◽  
Ferenc Olajos ◽  
...  

AbstractClinical guidelines for decision-making in chronic kidney disease (CKD) consider parathormone (PTH) levels. The measured PTH values differ if novel full length PTH(1-84) assays are used instead of earlier intact iPTH assays. In this study we analyzed how the classification of CKD patients alters when iPTH assays are switched to PTH(1-84) assays.Plasma samples were collected prior to dialysis sessions from 110 consecutive CKD patients on maintenance hemodialysis. PTH levels were determined with iPTH assays (Elecsys, Architect and DiaSorin Liaison N-tact) and PTH(1-84) assays (Elecsys and Liaison). Using KDIGO guidelines patients were classified as being below, above and in the recommended target range (RTR) of PTH. The results of classification with different assays were evaluated and, a novel calculation method of RTR was implemented.The prevalence of patients with PTH in RTR is comparable with each assay, but the individual patients differed. PTH(1-84) Elecsys and Liaison assays classified more patients as being below RTR than iPTH Elecsys and Architect but not Liaison N-tact assay (27.3%, 22.7% vs. 41%, 31.8%, and 36.4%, respectively). In turn, PTH(1-84) Elecsys and Liaison assays identified less CKD patients with PTH above the RTR than iPTH except N-tact assays (6.4%, 10% vs. 16.3%, 19%, and 6.3%, respectively). Using our calculation method, our discrimination values for PTH(1-84) assays to achieve classification identical to that with iPTH Elecsys were lower than those recommended by the manufacturer.Current guidelines for the treatment of secondary hyperparathyroidism in CKD should consider the type of assays used for PTH measurement. Each laboratory should assess its own RTR for PTH tests to achieve comparable classification. The presented calculation is simple, it mimics an everyday situation, switching from one assay to another one, and provides useful RTR values for PTH tests.


Author(s):  
L. A. Pesotskaia ◽  
N. V. Hlukhova ◽  
T. O. Tretiak ◽  
O. V. Pysarevskaia ◽  
M. H. Hetman ◽  
...  

There were 52 healthy people. Of them, 28 were in junior courses and 24 were in senior courses. According to the results of psychological tests there were identified three types of thinking: visual-figurative and verbal-logical, intuitive. Kirlian photographs of the surveyed individuals fingers in the experimental device «REC 1» were done. The method of acquiring, processing, and subsequent analysis included analog-to-digital conversion Kirlian images to highlight areas of illumination of the individual fingers, binarization of images with the threshold brightness, calculation of the area of corona glow. The obtained results complemented results of psychological tests, revealed mixed types of thinking, their expression, the dependence of the period of study at the university. To identify the emotional characteristics of the students associated with the constitutional type of the individual, selectively kronograf research on Polarod color photographic film with computer processing of scanned images carried out. Check radiation around the fingers of people surveyed in the color film, determination of the ratios of colors in the image, their energy allowed to determine the type of emotional activation of student body's potential abilities that affects motivation and optimize learning at the university.


2013 ◽  
Vol 4 (1) ◽  
Author(s):  
Lisa R Marr-Lyon ◽  
Gireesh V Gupchup ◽  
Joe R Anderson

Introduction: Since the development of the 10 item Purdue Pharmacist Directive Guidance (PPDG) Scale several studies of the psychometric properties of the PPDG have been conducted. Although Cronbach's alpha was calculated as a means of internal consistency reliability, a demonstration of the mean centering of the individual items from the instrument were not explored. Objectives: This study focused on investigating the mean stabilization of items within the PPDG as they pertain to Cronbach's reliability coefficient calculation. Methods: Using item analysis procedures in SPSS, the mean stability of items within the general factor of directive guidance and subscales of instruction and feedback and goal setting were examined for the PPDG. Results: Mean stability scores for entire PPDG scale and the subscales of instruction and feedback and goal setting were strong. Also, corrected item-total correlations and Cronbach's alphas following item deletion were good for the overall PPDG scale and the subscales. Conclusions: The results provide evidence to enhance understanding of the psychometric stability of the PPDG scale and its subscales.   Type: Original Research


2020 ◽  
Author(s):  
Luuk P.H. van de Rijt ◽  
A. John van Opstal ◽  
Marc M. van Wanrooij

AbstractThe cochlear implant (CI) allows profoundly deaf individuals to partially recover hearing. Still, due to the coarse acoustic information provided by the implant, CI users have considerable difficulties in recognizing speech, especially in noisy environments, even years after implantation. CI users therefore rely heavily on visual cues to augment speech comprehension, more so than normal-hearing individuals. However, it is unknown how attention to one (focused) or both (divided) modalities plays a role in multisensory speech recognition. Here we show that unisensory speech listening and speech reading were negatively impacted in divided-attention tasks for CI users - but not for normal-hearing individuals. Our psychophysical experiments revealed that, as expected, listening thresholds were consistently better for the normal-hearing, while lipreading thresholds were largely similar for the two groups. Moreover, audiovisual speech recognition for normal-hearing individuals could be described well by probabilistic summation of auditory and visual speech recognition, while CI users were better integrators than expected from statistical facilitation alone. Our results suggest that this benefit in integration, however, comes at a cost. Unisensory speech recognition is degraded for CI users when attention needs to be divided across modalities, i.e. in situations with uncertainty about the upcoming stimulus modality. We conjecture that CI users exhibit an integration-attention trade-off. They focus solely on a single modality during focused-attention tasks, but need to divide their limited attentional resources to more modalities during divided-attention tasks. We argue that in order to determine the benefit of a CI for speech comprehension, situational factors need to be discounted by presenting speech in realistic or complex audiovisual environments.Significance statementDeaf individuals using a cochlear implant require significant amounts of effort to listen in noisy environments due to their impoverished hearing. Lipreading can benefit them and reduce the burden of listening by providing an additional source of information. Here we show that the improved speech recognition for audiovisual stimulation comes at a cost, however, as the cochlear-implant users now need to listen and speech-read simultaneously, paying attention to both modalities. The data suggests that cochlear-implant users run into the limits of their attentional resources, and we argue that they, unlike normal-hearing individuals, always need to consider whether a multisensory benefit outweighs the unisensory cost in everyday environments.


Author(s):  
Ismail Burud ◽  
Kavitha Nagandla ◽  
Puneet Agarwal

Background: Item analysis is a quality assurance of examining the performance of the individual test items that measures the validity and reliability of exams. This study was performed to evaluate the quality of the test items with respect to their performance on difficulty index (DFI), Discriminatory index (DI) and assessment of functional and non-functional distractors (FD and NFD).Methods: This study was performed on the summative examination undertaken by 113 students. The analyses include 120 one best answers (OBAs) and 360 distractors.Results: Out of the 360 distractors, 85 distractors were chosen by less than 5% with the distractor efficiency of 23.6%. About 47 (13%) items had no NFDs while 51 (14%), 30 (8.3%), and 4 (1.1%) items contained 1, 2, and 3 NFDs respectively. Majority of the items showed excellent difficulty index (50.4%, n=42) and fair discrimination (37%, n=33). The questions with excellent difficulty index and discriminatory index showed statistical significance with 1NFD and 2 NFD (p=0.03).Conclusions: The post evaluation of item performance in any exam in one of the quality assurance method of identifying the best performing item for quality question bank. The distractor efficiency gives information on the overall quality of item.


2020 ◽  
Vol 117 (29) ◽  
pp. 16920-16927 ◽  
Author(s):  
John Plass ◽  
David Brang ◽  
Satoru Suzuki ◽  
Marcia Grabowecky

Visual speech facilitates auditory speech perception, but the visual cues responsible for these benefits and the information they provide remain unclear. Low-level models emphasize basic temporal cues provided by mouth movements, but these impoverished signals may not fully account for the richness of auditory information provided by visual speech. High-level models posit interactions among abstract categorical (i.e., phonemes/visemes) or amodal (e.g., articulatory) speech representations, but require lossy remapping of speech signals onto abstracted representations. Because visible articulators shape the spectral content of speech, we hypothesized that the perceptual system might exploit natural correlations between midlevel visual (oral deformations) and auditory speech features (frequency modulations) to extract detailed spectrotemporal information from visual speech without employing high-level abstractions. Consistent with this hypothesis, we found that the time–frequency dynamics of oral resonances (formants) could be predicted with unexpectedly high precision from the changing shape of the mouth during speech. When isolated from other speech cues, speech-based shape deformations improved perceptual sensitivity for corresponding frequency modulations, suggesting that listeners could exploit this cross-modal correspondence to facilitate perception. To test whether this type of correspondence could improve speech comprehension, we selectively degraded the spectral or temporal dimensions of auditory sentence spectrograms to assess how well visual speech facilitated comprehension under each degradation condition. Visual speech produced drastically larger enhancements during spectral degradation, suggesting a condition-specific facilitation effect driven by cross-modal recovery of auditory speech spectra. The perceptual system may therefore use audiovisual correlations rooted in oral acoustics to extract detailed spectrotemporal information from visual speech.


1987 ◽  
Vol 30 (4) ◽  
pp. 480-493 ◽  
Author(s):  
Laurel J. Dent ◽  
F. Blair Simmons ◽  
Robert L. White ◽  
Robert A. Roberts

Four profoundly deaf adults, each a recent recipient of a scala tympani implant, underwent auditory and auditory-visual speech comprehension evaluations. Two subjects had multiple-electrode auditory prostheses, and 2 had single-electrode implants. All subjects were tested preoperatively with a high-power hearing aid, and postoperatively with a single-channel wearable sound processor. Reported here are the results of the first formal speech recognition tests which were conducted during the 8 months after the sound processor fitting. Three subjects had used the single-channel processor on a daily basis for up to 8 months at the time of postoperative testing. The 4th subject was a nonuser. On listening tests, a comparison between pre- and post-implant scores revealed little difference for any subject. On postoperative speechreading tasks, all subjects identified medial consonant phonemes and 2-digit numerals better with stimulation than without. The 3 frequent users of the device experienced significant improvement on connected-discourse tracking, and their speechreading of videotaped and live voice CID Everyday Sentences (Davis & Silverman, 1978) was enhanced with the addition of stimulation. The nonuser was a very proficient speechreader at the outset and exhibited no significant difference on connected-discourse tracking with and without stimulation. Moreover her ability to speechread Everyday Sentences was hampered slightly by the addition of stimulation. This single-channel sound processor functioned as a sensory supplement for the 3 frequent users, but no subject was able to use the processor as a sensory substitute.


Sign in / Sign up

Export Citation Format

Share Document