Facial Affect Recognition in Myasthenia Gravis

2013 ◽  
Vol 16 ◽  
Author(s):  
Esther Lázaro ◽  
Imanol Amayra ◽  
Juan Francisco López-Paz ◽  
Amaia Jometón ◽  
Natalia Martín ◽  
...  

AbstractThe assessment of facial expression is an important aspect of a clinical neurological examination, both as an indicator of a mood disorder and as a sign of neurological damage. To date, although studies have been conducted on certain psychosocial aspects of myasthenia, such as quality of life and anxiety, and on neuropsychological aspects such as memory, no studies have directly assessed facial emotion recognition accuracy. The aim of this study was to assess the facial emotion recognition accuracy (fear, surprise, sadness, happiness, anger, and disgust), empathy, and reaction time of patients with myasthenia. Thirty-five patients with myasthenia and 36 healthy controls were tested for their ability to differentiate emotional facial expressions. Participants were matched with respect to age, gender, and education level. Their ability to differentiate emotional facial expressions was evaluated using the computer-based program Feel Test. The data showed that myasthenic patients scored significantly lower (p < 0.05) than healthy controls in the total Feel score, fear, surprise, and higher reaction time. The findings suggest that the ability to recognize facial affect may be reduced in individuals with myasthenia.

2017 ◽  
Vol 29 (5) ◽  
pp. 1749-1761 ◽  
Author(s):  
Johanna Bick ◽  
Rhiannon Luyster ◽  
Nathan A. Fox ◽  
Charles H. Zeanah ◽  
Charles A. Nelson

AbstractWe examined facial emotion recognition in 12-year-olds in a longitudinally followed sample of children with and without exposure to early life psychosocial deprivation (institutional care). Half of the institutionally reared children were randomized into foster care homes during the first years of life. Facial emotion recognition was examined in a behavioral task using morphed images. This same task had been administered when children were 8 years old. Neutral facial expressions were morphed with happy, sad, angry, and fearful emotional facial expressions, and children were asked to identify the emotion of each face, which varied in intensity. Consistent with our previous report, we show that some areas of emotion processing, involving the recognition of happy and fearful faces, are affected by early deprivation, whereas other areas, involving the recognition of sad and angry faces, appear to be unaffected. We also show that early intervention can have a lasting positive impact, normalizing developmental trajectories of processing negative emotions (fear) into the late childhood/preadolescent period.


2020 ◽  
Author(s):  
Nazire Duran ◽  
ANTHONY P. ATKINSON

Certain facial features provide useful information for recognition of facial expressions. In two experiments, we investigated whether foveating informative features of briefly presented expressions improves recognition accuracy and whether these features are targeted reflexively when not foveated. Angry, fearful, surprised, and sad or disgusted expressions were presented briefly at locations which would ensure foveation of specific features. Foveating the mouth of fearful, surprised and disgusted expressions improved emotion recognition compared to foveating an eye or cheek or the central brow. Foveating the brow lead to equivocal results in anger recognition across the two experiments, which might be due to the different combination of emotions used. There was no consistent evidence suggesting that reflexive first saccades targeted emotion-relevant features; instead, they targeted the closest feature to initial fixation. In a third experiment, angry, fearful, surprised and disgusted expressions were presented for 5 seconds. Duration of task-related fixations in the eyes, brow, nose and mouth regions was modulated by the presented expression. Moreover, longer fixation at the mouth positively correlated with anger and disgust accuracy both when these expressions were freely viewed (Experiment 3) and when briefly presented at the mouth (Experiment 2). Finally, an overall preference to fixate the mouth across all expressions correlated positively with anger and disgust accuracy. These findings suggest that foveal processing of informative features is functional/contributory to emotion recognition, but they are not automatically sought out when not foveated, and that facial emotion recognition performance is related to idiosyncratic gaze behaviour.


PLoS ONE ◽  
2021 ◽  
Vol 16 (12) ◽  
pp. e0260814
Author(s):  
Nazire Duran ◽  
Anthony P. Atkinson

Certain facial features provide useful information for recognition of facial expressions. In two experiments, we investigated whether foveating informative features of briefly presented expressions improves recognition accuracy and whether these features are targeted reflexively when not foveated. Angry, fearful, surprised, and sad or disgusted expressions were presented briefly at locations which would ensure foveation of specific features. Foveating the mouth of fearful, surprised and disgusted expressions improved emotion recognition compared to foveating an eye or cheek or the central brow. Foveating the brow led to equivocal results in anger recognition across the two experiments, which might be due to the different combination of emotions used. There was no consistent evidence suggesting that reflexive first saccades targeted emotion-relevant features; instead, they targeted the closest feature to initial fixation. In a third experiment, angry, fearful, surprised and disgusted expressions were presented for 5 seconds. Duration of task-related fixations in the eyes, brow, nose and mouth regions was modulated by the presented expression. Moreover, longer fixation at the mouth positively correlated with anger and disgust accuracy both when these expressions were freely viewed (Experiment 2b) and when briefly presented at the mouth (Experiment 2a). Finally, an overall preference to fixate the mouth across all expressions correlated positively with anger and disgust accuracy. These findings suggest that foveal processing of informative features is functional/contributory to emotion recognition, but they are not automatically sought out when not foveated, and that facial emotion recognition performance is related to idiosyncratic gaze behaviour.


2019 ◽  
Vol 34 (2) ◽  
pp. 254-263
Author(s):  
Gill Terrett ◽  
Kimberly Mercuri ◽  
Elizabeth Pizarro-Campagna ◽  
Laila Hugrass ◽  
H Valerie Curran ◽  
...  

Background: Long-term opiate users experience pervasive social difficulties, but there has been surprisingly limited research focused on social-cognitive functioning in this population. Aim: The aim of this study was to investigate whether three important aspects of social cognition (facial emotion recognition, theory of mind (ToM) and rapid facial mimicry) differ between long-term opiate users and healthy controls. Methods: The participants were 25 long-term opiate users who were enrolled in opiate substitution programmes, and 25 healthy controls. Facial emotion recognition accuracy was indexed by responses to 60 photographs of faces depicting the six basic emotions (happiness, sadness, anger, fear, surprise and disgust). ToM was assessed using the Reading the Mind in the Eyes task, which requires participants to infer mental states of others from partial facial cues. Rapid facial mimicry was assessed by recording activity in the zygomaticus major and corrugator supercilii muscle regions while participants passively viewed images of happy and angry facial expressions. Results: Relative to the control group, the opiate user group exhibited deficits in both facial emotion recognition and ToM. Moreover, only control participants exhibited typical rapid facial mimicry responses to happy facial expressions. Conclusions: These data indicate that long-term opiate users exhibit abnormalities in three distinct areas of social-cognitive processing, pointing to the need for additional work to establish how social-cognitive functioning relates to functional outcomes in this group. Such work may ultimately inform the development of interventions aimed at improving treatment outcomes for long-term opiate users.


2021 ◽  
Author(s):  
Michael J. Spilka ◽  
William R. Keller ◽  
Robert W. Buchanan ◽  
james gold ◽  
James I. Koenig ◽  
...  

Objective: Difficulties in social cognition are common in individuals with schizophrenia (SZ) and are not ameliorated by antipsychotic treatment. Intranasal oxytocin (OT) administration has been explored as a potential intervention to improve social cognition; however, results are inconsistent, suggesting potential individual difference variables that may influence treatment response. Less is known about the relationship between endogenous OT and social cognition in SZ, knowledge of which may improve the development of OT-focused therapies. We examined plasma OT in relationship to facial emotion recognition and visual attention to salient facial features in SZ and controls. Methods: Forty-two individuals with SZ and 23 healthy controls viewed photographs of facial expressions of varying emotional intensity and identified the emotional expression displayed. Participants’ gaze behavior during the task was recorded via eye tracking. Plasma oxytocin concentrations were determined by radioimmunoassay. Results: SZ were less accurate than controls at identifying high intensity fearful facial expressions and low intensity sad expressions. Lower facial emotion recognition accuracy was associated with lower plasma OT levels in SZ but not controls. SZ had reduced visual attention to the nose region compared to controls; however, OT was not associated with gaze behavior. Conclusion: Individual differences in endogenous OT predict facial emotion recognition ability in SZ but are not associated with visual attention to salient facial features. Increased understanding of the association between endogenous OT and social cognitive abilities in SZ may help improve the design and interpretation of OT-focused clinical trials in SZ.


2019 ◽  
Vol 123 (4) ◽  
pp. 1099-1116
Author(s):  
Milica Mitrovic ◽  
Milica Ristic ◽  
Bojana Dimitrijevic ◽  
Marina Hadzi Pesic

The recognition of facial signals has a crucial role in social interaction. It is well known that people suffering from paranoid schizophrenia have problems in the social domain, predominantly related to misinterpreting the intentions, emotions, and actions of others. The aim of this study was to examine whether there are differences in facial emotion recognition between people with paranoid schizophrenia and healthy controls. In addition, we examined the correlation between facial emotion recognition and the expression of persecutory ideation in people suffering from paranoid schizophrenia. The study involved 60 participants, 30 of whom suffered from paranoid schizophrenia and 30 healthy controls, equalized by gender, age, and education. The following instruments were used: Japanese and Caucasian Facial Expressions of Emotion and Neutral Faces and the Persecutory Ideation Questionnaire. Compared with the controls, people suffering from paranoid schizophrenia were significantly less accurate in recognizing the following emotions: surprise, contempt, sadness, disgust, and emotionally neutral faces. Since the attribution of emotions to emotionally neutral faces is an important finding that could be linked with the social (dis)functionality of people suffering from paranoid schizophrenia, we analyzed and compared the wrong answers given by the two groups and found some differences between them. The results show that persecutory ideation has a statistically significant negative correlation with the successful recognition of emotionally neutral faces. All of the findings lead to the conclusion that paranoid schizophrenia, and within it the existence of persecutory ideation, leads to problems in recognizing the basic facial signals that form the foundation of everyday social interaction.


2016 ◽  
Vol 23 (1) ◽  
pp. 23-33 ◽  
Author(s):  
Kelly Hubble ◽  
Katie Daughters ◽  
Antony S.R. Manstead ◽  
Aled Rees ◽  
Anita Thapar ◽  
...  

AbstractObjectives: Previous studies have found that oxytocin (OXT) can improve the recognition of emotional facial expressions; it has been proposed that this effect is mediated by an increase in attention to the eye-region of faces. Nevertheless, evidence in support of this claim is inconsistent, and few studies have directly tested the effect of oxytocin on emotion recognition via altered eye-gaze Methods: In a double-blind, within-subjects, randomized control experiment, 40 healthy male participants received 24 IU intranasal OXT and placebo in two identical experimental sessions separated by a 2-week interval. Visual attention to the eye-region was assessed on both occasions while participants completed a static facial emotion recognition task using medium intensity facial expressions. Results: Although OXT had no effect on emotion recognition accuracy, recognition performance was improved because face processing was faster across emotions under the influence of OXT. This effect was marginally significant (p<.06). Consistent with a previous study using dynamic stimuli, OXT had no effect on eye-gaze patterns when viewing static emotional faces and this was not related to recognition accuracy or face processing time. Conclusions: These findings suggest that OXT-induced enhanced facial emotion recognition is not necessarily mediated by an increase in attention to the eye-region of faces, as previously assumed. We discuss several methodological issues which may explain discrepant findings and suggest the effect of OXT on visual attention may differ depending on task requirements. (JINS, 2017, 23, 23–33)


2021 ◽  
pp. 1-10
Author(s):  
Daniel T. Burley ◽  
Christopher W. Hobson ◽  
Dolapo Adegboye ◽  
Katherine H. Shelton ◽  
Stephanie H.M. van Goozen

Abstract Impaired facial emotion recognition is a transdiagnostic risk factor for a range of psychiatric disorders. Childhood behavioral difficulties and parental emotional environment have been independently associated with impaired emotion recognition; however, no study has examined the contribution of these factors in conjunction. We measured recognition of negative (sad, fear, anger), neutral, and happy facial expressions in 135 children aged 5–7 years referred by their teachers for behavioral problems. Parental emotional environment was assessed for parental expressed emotion (EE) – characterized by negative comments, reduced positive comments, low warmth, and negativity towards their child – using the 5-minute speech sample. Child behavioral problems were measured using the teacher-informant Strengths and Difficulties Questionnaire (SDQ). Child behavioral problems and parental EE were independently associated with impaired recognition of negative facial expressions specifically. An interactive effect revealed that the combination of both factors was associated with the greatest risk for impaired recognition of negative faces, and in particular sad facial expressions. No relationships emerged for the identification of happy facial expressions. This study furthers our understanding of multidimensional processes associated with the development of facial emotion recognition and supports the importance of early interventions that target this domain.


2016 ◽  
Vol 28 (7) ◽  
pp. 1165-1179 ◽  
Author(s):  
J. Pietschnig ◽  
L. Schröder ◽  
I. Ratheiser ◽  
I. Kryspin-Exner ◽  
M. Pflüger ◽  
...  

ABSTRACTBackground:Impairments in facial emotion recognition (FER) have been detected in patients with Parkinson disease (PD). Presently, we aim at assessing differences in emotion recognition performance in PD patient groups with and without mild forms of cognitive impairment (MCI) compared to healthy controls.Methods:Performance on a concise emotion recognition test battery (VERT-K) of three groups of 97 PD patients was compared with an age-equivalent sample of 168 healthy controls. Patients were categorized into groups according to two well-established classifications of MCI according to Petersen's (cognitively intact vs. amnestic MCI, aMCI, vs. non-amnestic MCI, non-aMCI) and Litvan's (cognitively intact vs. single-domain MCI, sMCI, vs. multi-domain MCI, mMCI) criteria. Patients and controls underwent individual assessments using a comprehensive neuropsychological test battery examining attention, executive functioning, language, and memory (Neuropsychological Test Battery Vienna, NTBV), the Beck Depression Inventory, and a measure of premorbid IQ (WST).Results:Cognitively intact PD patients and patients with MCI in PD (PD-MCI) showed significantly worse emotion recognition performance when compared to healthy controls. Between-groups effect sizes were substantial, showing non-trivial effects in all comparisons (Cohen's ds from 0.31 to 1.22). Moreover, emotion recognition performance was higher in women, positively associated with premorbid IQ and negatively associated with age. Depressive symptoms were not related to FER.Conclusions:The present investigation yields further evidence for impaired FER in PD. Interestingly, our data suggest FER deficits even in cognitively intact PD patients indicating FER dysfunction prior to the development of overt cognitive dysfunction. Age showed a negative association whereas IQ showed a positive association with FER.


2021 ◽  
Vol 12 ◽  
Author(s):  
Paula J. Webster ◽  
Shuo Wang ◽  
Xin Li

Different styles of social interaction are one of the core characteristics of autism spectrum disorder (ASD). Social differences among individuals with ASD often include difficulty in discerning the emotions of neurotypical people based on their facial expressions. This review first covers the rich body of literature studying differences in facial emotion recognition (FER) in those with ASD, including behavioral studies and neurological findings. In particular, we highlight subtle emotion recognition and various factors related to inconsistent findings in behavioral studies of FER in ASD. Then, we discuss the dual problem of FER – namely facial emotion expression (FEE) or the production of facial expressions of emotion. Despite being less studied, social interaction involves both the ability to recognize emotions and to produce appropriate facial expressions. How others perceive facial expressions of emotion in those with ASD has remained an under-researched area. Finally, we propose a method for teaching FER [FER teaching hierarchy (FERTH)] based on recent research investigating FER in ASD, considering the use of posed vs. genuine emotions and static vs. dynamic stimuli. We also propose two possible teaching approaches: (1) a standard method of teaching progressively from simple drawings and cartoon characters to more complex audio-visual video clips of genuine human expressions of emotion with context clues or (2) teaching in a field of images that includes posed and genuine emotions to improve generalizability before progressing to more complex audio-visual stimuli. Lastly, we advocate for autism interventionists to use FER stimuli developed primarily for research purposes to facilitate the incorporation of well-controlled stimuli to teach FER and bridge the gap between intervention and research in this area.


Sign in / Sign up

Export Citation Format

Share Document