Anticipatory eye movements and long-term memory in early infancy

2016 ◽  
Vol 58 (7) ◽  
pp. 841-851 ◽  
Author(s):  
Audrey M.B. Wong-Kee-You ◽  
Scott A. Adler
Author(s):  
Kai Essig ◽  
Oleg Strogan ◽  
Helge Ritter ◽  
Thomas Schack

Various computational models of visual attention rely on the extraction of salient points or proto-objects, i.e., discrete units of attention, computed from bottom-up image features. In recent years, different solutions integrating top-down mechanisms were implemented, as research has shown that although eye movements initially are solely influenced by bottom-up information, after some time goal driven (high-level) processes dominate the guidance of visual attention towards regions of interest (Hwang, Higgins & Pomplun, 2009). However, even these improved modeling approaches are unlikely to generalize to a broader range of application contexts, because basic principles of visual attention, such as cognitive control, learning and expertise, have thus far not sufficiently been taken into account (Tatler, Hayhoe, Land & Ballard, 2011). In some recent work, the authors showed the functional role and representational nature of long-term memory structures for human perceptual skills and motor control. Based on these findings, the chapter extends a widely applied saliency-based model of visual attention (Walther & Koch, 2006) in two ways: first, it computes the saliency map using the cognitive visual attention approach (CVA) that shows a correspondence between regions of high saliency values and regions of visual interest indicated by participants’ eye movements (Oyekoya & Stentiford, 2004). Second, it adds an expertise-based component (Schack, 2012) to represent the influence of the quality of mental representation structures in long-term memory (LTM) and the roles of learning on the visual perception of objects, events, and motor actions.


1980 ◽  
Vol 51 (3) ◽  
pp. 700 ◽  
Author(s):  
Vicki K. Mast ◽  
Jeffrey W. Fagen ◽  
Carolyn K. Rovee-Collier ◽  
Margaret W. Sullivan

2020 ◽  
Author(s):  
James E. Kragel ◽  
Stephan Schuele ◽  
Stephen VanHaerents ◽  
Joshua M. Rosenow ◽  
Joel L. Voss

Although the human hippocampus is necessary for long-term memory, controversial findings suggest that hippocampal computations support short-term memory in the service of guiding effective behaviors during learning. We tested the counterintuitive theory that the hippocampus contributes to long-term memory through remarkably short-term processing, as reflected in the sequence of eye movements during encoding of naturalistic scenes. While viewing scenes for the first time, participants generated patterns of eye movements that reflected a shift from stimulus-driven to memory-driven viewing and signaled effective spatiotemporal memory formation. Hippocampal theta oscillations recorded from depth electrodes predicted this viewing pattern. Moreover, effective viewing patterns were preceded by shifts towards top-down influence of hippocampal theta on activity within cortical networks that support visual perception and visuospatial attention. The hippocampus thus supports short-term memory processing that coordinates perception, attention, and behavior in the service of effective spatiotemporal learning. These findings motivate re-interpretation of long-term memory disorders as reflecting loss of the organizing influence of hippocampal short-term memory on learning.


2019 ◽  
Vol 19 (10) ◽  
pp. 149a
Author(s):  
Lisa F Schwetlick ◽  
Hans A Trukenbrod ◽  
Ralf Engbert

NeuroImage ◽  
2018 ◽  
Vol 179 ◽  
pp. 348-356 ◽  
Author(s):  
Denis Pélisson ◽  
Ouazna Habchi ◽  
Muriel T.N. Panouillères ◽  
Charles Hernoux ◽  
Alessandro Farnè

Author(s):  
Seana Coulson ◽  
Thomas P Urbach ◽  
Marta Kutas

AbstractWe describe the space structuring model, a model of language comprehension inspired by ideas in cognitive linguistics, focusing on its capacity to explain the sorts of inferences needed to understand one-line jokes. One process posited in the model is frame-shifting, the semantic and pragmatic reanalysis in which elements of the existing message-level representation are mapped into a new frame retrieved from long-term memory. To test this model, we recorded participants' eye movements with a headband-mounted eye-tracker while they read sentences that ended either as a joke, or as nonfunny controls (“She read so much about the bad effects of smoking she decided to give up the reading/habit.”) . Only jokes required frame-shifting; nonjoke endings were consistent with the contextually evoked frame. Though initial gaze durations were the same for jokes and non-jokes, total viewing duration was longer for the jokes and participants were more likely to make regressive (leftward) eye movements after reading the “punch word” of a joke. Results are consistent with the psychological reality of some process like frame-shifting, suggesting readers literally revisit aspects of the prior context while apprehending one-line jokes.


Sign in / Sign up

Export Citation Format

Share Document