Visual Recognition of Mother by Young Infants: Facilitation by Speech

Perception ◽  
1993 ◽  
Vol 22 (10) ◽  
pp. 1133-1153 ◽  
Author(s):  
Denis Burnham

Infants recognise their mother's voice at birth but appear not to recognise visual-only presentations of her face until around 3 months. In a series of experiments visual discrimination by infants aged 1, 3, and 5 months of their mother's and a female stranger's face was investigated in visual-only and visual-plus-speech conditions. In the first experiment these infants' discrimination of mother's and female stranger's faces was measured by their visual-fixation-preference scores. Discrimination was found to be facilitated by the addition of speech information. In experiment 2 naive adults viewed silent videotapes of infants from experiment 1 and judged whether the mother had been presented on the infants' left or right. This added further information to the fixation-preference results of experiment 1: it was found that 1-month-olds discriminate mother's and stranger's face only in the presence of speech information, whereas 3-month-olds also do so in visual-only conditions. In experiments 3 and 4 the relative salience of lip movements and voice information in visual recognition of mother's face was investigated. In experiment 3, no significant differences in infants' visual-fixation-preference scores were obtained. However, in experiment 4 adults' ‘where is mother?’ judgments of videotapes from experiment 3 were found to be more accurate in the voice than in the lip-movements conditions, especially for the 3-month-olds and more accurate when mother rather than stranger was talking. It is concluded that young infants' visual recognition of mother is facilitated by addition of speech information, that it is primarily the voice component of speech that causes this facilitation, and that social discrimination is best indexed by a dependent variable which is sensitive to a range of facial cues provided by infants.

Author(s):  
Weigao Su ◽  
Daibo Liu ◽  
Taiyuan Zhang ◽  
Hongbo Jiang

Motion sensors in modern smartphones have been exploited for audio eavesdropping in loudspeaker mode due to their sensitivity to vibrations. In this paper, we further move one step forward to explore the feasibility of using built-in accelerometer to eavesdrop on the telephone conversation of caller/callee who takes the phone against cheek-ear and design our attack Vibphone. The inspiration behind Vibphone is that the speech-induced vibrations (SIV) can be transmitted through the physical contact of phone-cheek to accelerometer with the traces of voice content. To this end, Vibphone faces three main challenges: i) Accurately detecting SIV signals from miscellaneous disturbance; ii) Combating the impact of device diversity to work with a variety of attack scenarios; and iii) Enhancing feature-agnostic recognition model to generalize to newly issued devices and reduce training overhead. To address these challenges, we first conduct an in-depth investigation on SIV features to figure out the root cause of device diversity impacts and identify a set of critical features that are highly relevant to the voice content retained in SIV signals and independent of specific devices. On top of these pivotal observations, we propose a combo method that is the integration of extracted critical features and deep neural network to recognize speech information from the spectrogram representation of acceleration signals. We implement the attack using commodity smartphones and the results show it is highly effective. Our work brings to light a fundamental design vulnerability in the vast majority of currently deployed smartphones, which may put people's speech privacy at risk during phone calls. We also propose a practical and effective defense solution. We validate that it is feasible to prevent audio eavesdropping by using random variation of sampling rate.


Author(s):  
Agnese Sile

When it comes to depicting ill or disabled children, the ethics of representation becomes increasingly complex. The perception of photographs as voyeuristic and objectifying is of particular concern here and resonates with widespread fear about the eroticisation, mistreatment and exploitation of children. Although these fears are reasonable, this view does not take into account the voice and agenda of the photographic subject, disregards the possibility of recognition and the participatory nature of photography. In this article, I focus on photography as a collaborative practice. I analyse two photographic projects by photographers/mothers that document their ill and dying daughters – Lesley McIntyre’s photographic essay The Time of Her Life (2004) and Elisabeth Zahnd Legnazzi’s Chiara A Journey Into Light (2009). Illness in these projects is not experienced in isolation. Instead, the photographs and accompanying texts provide a space to engage in a dialogue which is built on the interdependency of all the participants of the photographic act – the photographer, the subject of the photograph and the viewer. My aim is to question how these projects construct experiences and articulate private expressions of illness and how the photographs enhance and/or challenge the mother–daughter bond. Alan Radley’s critical analysis of representations of illness, Emmanuel Lévinas’s and Maurice Blanchot’s perspectives on ethical philosophy and visual social semiotics approach developed by Kress and Van Leeuwen provide a guiding framework for this study.


Perception ◽  
1978 ◽  
Vol 7 (5) ◽  
pp. 491-497 ◽  
Author(s):  
Jacques Mehler ◽  
Josiane Bertoncini ◽  
Michele Barriere ◽  
Dora Jassik-Gerschenfeld

Each of a group of one-month-old infants was reinforced, contingent upon nonnutritive sucking, with its mother's voice and the voice of a stranger. In this experiment, two conditions were applied. Under the first, the mother's speech was aimed at communicating with the infant, while, under the second, the mother's speech lacked prosodic and intonational aspects of normal speech. It was shown that infants will suck more for their mother's voices under the intonated condition only. It was concluded that a young infant prefers its own mother's voice provided the mother speaks normally.


2010 ◽  
Vol 139-141 ◽  
pp. 2154-2157
Author(s):  
Ji Xiang Lu ◽  
Ping Wang ◽  
Long Yi

The voice interaction in cockpit mainly includes speech recognition, enhancement and synthesis. This interaction transfers the speech information to the corresponding orders to make machines in cockpit work unmistaken, also feedback the execution results to users by speech output devices or some other ways. The speech enhancement technology is studied in this paper, aiming at the Voice Interactive. We propose an improved spectral subtraction (SS) algorithm based on auditory masking effect, by using two steps SS. The simulated results based on the segment SNR compared to the traditional SS show the effectiveness and superiority of the improved algorithm.


1996 ◽  
Vol 19 ◽  
pp. 515
Author(s):  
Derek Houston ◽  
Peter Jusczyk ◽  
Ann Marie Jusczyk

1983 ◽  
Vol 35 (2b) ◽  
pp. 169-194 ◽  
Author(s):  
Euan M. Macphail

Two series of experiments investigated short-term visual recognition memory in pigeons following lesions of the hyperstriatal complex; the first series used a choice technique, the second, a single-key go/no go technique. The results of the two series agreed, first, in finding impaired performance in hyperstriatal birds at long but not at short inter-trial intervals, and, second, in obtaining no evidence of differential rates of decay of traces in hyperstriatal and control subjects. A final experiment confirmed that the hyperstriatal birds were, as expected from previous work, impaired on reversals of colour and position discriminations. It is tentatively suggested that deficits following hyperstriatal damage in both recognition and reversal performance may be understood as being the consequence of an increased susceptibility to frustrating events in hyperstriatal subjects.


1997 ◽  
Vol 68 (6) ◽  
pp. 1041 ◽  
Author(s):  
John E. Richards ◽  
Theresa L. Gibson

Indoor and outdoor Navigation is a tough task for a visually impaired person and they would most of the time require human assistance. Existing solutions to this problem are in the form of smart canes and wearable. Both of which use sensors like on - board proximity and obstacle detection, as well as a haptic or auditory feedback system to warn the user of stationary or incoming obstacles so that they do not collide with any of them as they move. This approach has many drawbacks as it is not yet a stand - alone reliable device for the user to trust when navigating, and when frequently triggered in crowded areas, the feedback system will confuse the user with too many requests resulting in loss of actual information. Our Goal here is to create a Personalized assistant to the user, which they can interact naturally with their voice to mimic the aid of an actual human assistance while they are on the move. It works by using its object detection module with a high reliability training accuracy to detect the boundaries of objects in motion per frame and once the bounding box crosses the threshold accuracy, recognize the object in the box and pass the information to the system core, where it verifies if the information needed to be passed onto the user or not, if yes it passes the converted speech information to the voice interaction model. The voice interaction model is consent-based, it would accept and respond to navigation queries from the user and will intelligently inform them about the obstacle which needs to be avoided. This ensures only the essential information in the form of voice requests is passed onto the user, which they can use to navigate and also interact with the assistant for more information.


Sign in / Sign up

Export Citation Format

Share Document