Neural dynamics of grouping and segmentation explain properties of visual crowding.

2017 ◽  
Vol 124 (4) ◽  
pp. 483-504 ◽  
Author(s):  
Gregory Francis ◽  
Mauro Manassi ◽  
Michael H. Herzog
2019 ◽  
Author(s):  
Adrien Doerig ◽  
Lynn Schmittwilken ◽  
Bilge Sayim ◽  
Mauro Manassi ◽  
Michael H. Herzog

AbstractClassically, visual processing is described as a cascade of local feedforward computations. Feedforward Convolutional Neural Networks (ffCNNs) have shown how powerful such models can be. However, using visual crowding as a well-controlled challenge, we previously showed that no classic model of vision, including ffCNNs, can explain human global shape processing (1). Here, we show that Capsule Neural Networks (CapsNets; 2), combining ffCNNs with recurrent grouping and segmentation, solve this challenge. We also show that ffCNNs and standard recurrent CNNs do not, suggesting that the grouping and segmentation capabilities of CapsNets are crucial. Furthermore, we provide psychophysical evidence that grouping and segmentation are implemented recurrently in humans, and show that CapsNets reproduce these results well. We discuss why recurrence seems needed to implement grouping and segmentation efficiently. Together, we provide mutually reinforcing psychophysical and computational evidence that a recurrent grouping and segmentation process is essential to understand the visual system and create better models that harness global shape computations.Author SummaryFeedforward Convolutional Neural Networks (ffCNNs) have revolutionized computer vision and are deeply transforming neuroscience. However, ffCNNs only roughly mimic human vision. There is a rapidly expanding body of literature investigating differences between humans and ffCNNs. Several findings suggest that, unlike humans, ffCNNs rely mostly on local visual features. Furthermore, ffCNNs lack recurrent connections, which abound in the brain. Here, we use visual crowding, a well-known psychophysical phenomenon, to investigate recurrent computations in global shape processing. Previously, we showed that no model based on the classic feedforward framework of vision can explain global effects in crowding. Here, we show that Capsule Neural Networks (CapsNets), combining ffCNNs with recurrent grouping and segmentation, solve this challenge. ffCNNs and recurrent CNNs with lateral and top-down recurrent connections do not, suggesting that grouping and segmentation are crucial for human-like global computations. Based on these results, we hypothesize that one computational function of recurrence is to efficiently implement grouping and segmentation. We provide psychophysical evidence that, indeed, grouping and segmentation is based on time consuming recurrent processes in the human brain. CapsNets reproduce these results too. Together, we provide mutually reinforcing computational and psychophysical evidence that a recurrent grouping and segmentation process is essential to understand the visual system and create better models that harness global shape computations.


2020 ◽  
Author(s):  
Amandine Lassalle ◽  
Michael X Cohen ◽  
Laura Dekkers ◽  
Elizabeth Milne ◽  
Rasa Gulbinaite ◽  
...  

Background: People with an Autism Spectrum Condition diagnosis (ASD) are hypothesized to show atypical neural dynamics, reflecting differences in neural structure and function. However, previous results regarding neural dynamics in autistic individuals have not converged on a single pattern of differences. It is possible that the differences are cognitive-set-specific, and we therefore measured EEG in autistic individuals and matched controls during three different cognitive states: resting, visual perception, and cognitive control.Methods: Young adults with and without an ASD (N=17 in each group) matched on age (range 20 to 30 years), sex, and estimated Intelligence Quotient (IQ) were recruited. We measured their behavior and their EEG during rest, a task requiring low-level visual perception of gratings of varying spatial frequency, and the “Simon task” to elicit activity in the executive control network. We computed EEG power and Inter-Site Phase Clustering (ISPC; a measure of connectivity) in various frequency bands.Results: During rest, there were no ASD vs. controls differences in EEG power, suggesting typical oscillation power at baseline. During visual processing, without pre-baseline normalization, we found decreased broadband EEG power in ASD vs. controls, but this was not the case during the cognitive control task. Furthermore, the behavioral results of the cognitive control task suggest that autistic adults were better able to ignore irrelevant stimuli.Conclusions: Together, our results defy a simple explanation of overall differences between ASD and controls, and instead suggest a more nuanced pattern of altered neural dynamics that depend on which neural networks are engaged.


2017 ◽  
Vol 31 (3) ◽  
pp. 407-418 ◽  
Author(s):  
Gema Díaz-Blancat ◽  
Juan García-Prieto ◽  
Fernando Maestú ◽  
Francisco Barceló

2021 ◽  
Vol 3 (6) ◽  
Author(s):  
Arindam Singha ◽  
Anjan Kumar Ray ◽  
Arun Baran Samaddar
Keyword(s):  

A correction to this paper has been published: https://doi.org/10.1007/s42452-021-04606-4


2021 ◽  
Vol 12 (1) ◽  
Author(s):  
Hamidreza Abbaspourazad ◽  
Mahdi Choudhury ◽  
Yan T. Wong ◽  
Bijan Pesaran ◽  
Maryam M. Shanechi

AbstractMotor function depends on neural dynamics spanning multiple spatiotemporal scales of population activity, from spiking of neurons to larger-scale local field potentials (LFP). How multiple scales of low-dimensional population dynamics are related in control of movements remains unknown. Multiscale neural dynamics are especially important to study in naturalistic reach-and-grasp movements, which are relatively under-explored. We learn novel multiscale dynamical models for spike-LFP network activity in monkeys performing naturalistic reach-and-grasps. We show low-dimensional dynamics of spiking and LFP activity exhibited several principal modes, each with a unique decay-frequency characteristic. One principal mode dominantly predicted movements. Despite distinct principal modes existing at the two scales, this predictive mode was multiscale and shared between scales, and was shared across sessions and monkeys, yet did not simply replicate behavioral modes. Further, this multiscale mode’s decay-frequency explained behavior. We propose that multiscale, low-dimensional motor cortical state dynamics reflect the neural control of naturalistic reach-and-grasp behaviors.


Sign in / Sign up

Export Citation Format

Share Document