scholarly journals Locomotion induces stimulus-specific response enhancement in adult visual cortex

2017 ◽  
Author(s):  
Megumi Kaneko ◽  
Yu Fu ◽  
Michael P. Stryker

SUMMARYThe responses of neurons in the visual cortex (V1) of adult mammals have long been thought to be stable over long periods. Here, we investigated whether repeated exposure to specific stimuli would enhance V1 visual responses in mice using intrinsic signal imaging through the intact skull and two-photon imaging of calcium signals in single neurons. Mice ran on Styrofoam balls floating on air while viewing one of three different, high-contrast visual stimuli. V1 responses to the stimuli that were viewed by the animal were specifically enhanced, while responses to other stimuli were unaffected. Similar exposure in stationary mice, or in mice in which NMDA receptors were partially blocked, did not significantly enhance responses. These findings indicate that stimulus-specific plasticity in the adult visual cortex depends on concurrent locomotion, presumably as a result of the high-gain state of visual cortex induced by locomotion.Significance StatementWe report a rapid and persistent increase in visual cortical responses to visual stimuli presented during locomotion in intact mice. We first used a method that is completely non-invasive, intrinsic signal imaging through the intact skull. We then measured the same effects on single neurons using 2-photon calcium imaging and found that the increase in response to a particular stimulus produced by locomotion depends on how well the neuron is initially driven by the stimulus. To our knowledge, this is the first time such enhancement has been described in single neurons or using non-invasive measurements.

2019 ◽  
Vol 121 (6) ◽  
pp. 2202-2214 ◽  
Author(s):  
John P. McClure ◽  
Pierre-Olivier Polack

Multimodal sensory integration facilitates the generation of a unified and coherent perception of the environment. It is now well established that unimodal sensory perceptions, such as vision, are improved in multisensory contexts. Whereas multimodal integration is primarily performed by dedicated multisensory brain regions such as the association cortices or the superior colliculus, recent studies have shown that multisensory interactions also occur in primary sensory cortices. In particular, sounds were shown to modulate the responses of neurons located in layers 2/3 (L2/3) of the mouse primary visual cortex (V1). Yet, the net effect of sound modulation at the V1 population level remained unclear. In the present study, we performed two-photon calcium imaging in awake mice to compare the representation of the orientation and the direction of drifting gratings by V1 L2/3 neurons in unimodal (visual only) or multimodal (audiovisual) conditions. We found that sound modulation depended on the tuning properties (orientation and direction selectivity) and response amplitudes of V1 L2/3 neurons. Sounds potentiated the responses of neurons that were highly tuned to the cue’s orientation and direction but weakly active in the unimodal context, following the principle of inverse effectiveness of multimodal integration. Moreover, sound suppressed the responses of neurons untuned for the orientation and/or the direction of the visual cue. Altogether, sound modulation improved the representation of the orientation and direction of the visual stimulus in V1 L2/3. Namely, visual stimuli presented with auditory stimuli recruited a neuronal population better tuned to the visual stimulus orientation and direction than when presented alone. NEW & NOTEWORTHY The primary visual cortex (V1) receives direct inputs from the primary auditory cortex. Yet, the impact of sounds on visual processing in V1 remains controverted. We show that the modulation by pure tones of V1 visual responses depends on the orientation selectivity, direction selectivity, and response amplitudes of V1 neurons. Hence, audiovisual stimuli recruit a population of V1 neurons better tuned to the orientation and direction of the visual stimulus than unimodal visual stimuli.


Science ◽  
2019 ◽  
Vol 363 (6422) ◽  
pp. 64-69 ◽  
Author(s):  
Riccardo Beltramo ◽  
Massimo Scanziani

Visual responses in the cerebral cortex are believed to rely on the geniculate input to the primary visual cortex (V1). Indeed, V1 lesions substantially reduce visual responses throughout the cortex. Visual information enters the cortex also through the superior colliculus (SC), but the function of this input on visual responses in the cortex is less clear. SC lesions affect cortical visual responses less than V1 lesions, and no visual cortical area appears to entirely rely on SC inputs. We show that visual responses in a mouse lateral visual cortical area called the postrhinal cortex are independent of V1 and are abolished upon silencing of the SC. This area outperforms V1 in discriminating moving objects. We thus identify a collicular primary visual cortex that is independent of the geniculo-cortical pathway and is capable of motion discrimination.


Author(s):  
R. Oz ◽  
H. Edelman-Klapper ◽  
S. Nivinsky-Margalit ◽  
H. Slovin

AbstractIntra cortical microstimulation (ICMS) in the primary visual cortex (V1) can generate the visual perception of phosphenes and evoke saccades directed to the stimulated location in the retinotopic map. Although ICMS is widely used, little is known about the evoked spatio-temporal patterns of neural activity and their relation to neural responses evoked by visual stimuli or saccade generation. To investigate this, we combined ICMS with Voltage Sensitive Dye Imaging in V1 of behaving monkeys and measured neural activity at high spatial (meso-scale) and temporal resolution. Small visual stimuli and ICMS evoked population activity spreading over few mm that propagated to extrastriate areas. The population responses evoked by ICMS showed faster dynamics and different spatial propagation patterns. Neural activity was higher in trials w/saccades compared with trials w/o saccades. In conclusion, our results uncover the spatio-temporal patterns evoked by ICMS and their relation to visual processing and saccade generation.


2016 ◽  
Author(s):  
Dylan R Muir ◽  
Patricia Molina-Luna ◽  
Morgane M Roth ◽  
Fritjof Helmchen ◽  
Björn M Kampa

AbstractLocal excitatory connections in mouse primary visual cortex (V1) are stronger and more prevalent between neurons that share similar functional response features. However, the details of how functional rules for local connectivity shape neuronal responses in V1 remain unknown. We hypothesised that complex responses to visual stimuli may arise as a consequence of rules for selective excitatory connectivity within the local network in the superficial layers of mouse V1. In mouse V1 many neurons respond to overlapping grating stimuli (plaid stimuli) with highly selective and facilitatory responses, which are not simply predicted by responses to single gratings presented alone. This complexity is surprising, since excitatory neurons in V1 are considered to be mainly tuned to single preferred orientations. Here we examined the consequences for visual processing of two alternative connectivity schemes: in the first case, local connections are aligned with visual properties inherited from feedforward input (a ‘like-to-like’ scheme specifically connecting neurons that share similar preferred orientations); in the second case, local connections group neurons into excitatory subnetworks that combine and amplify multiple feedforward visual properties (a ‘feature binding’ scheme). By comparing predictions from large scale computational models with in vivo recordings of visual representations in mouse V1, we found that responses to plaid stimuli were best explained by a assuming ‘feature binding’ connectivity. Unlike under the ‘like-to-like’ scheme, selective amplification within feature-binding excitatory subnetworks replicated experimentally observed facilitatory responses to plaid stimuli; explained selective plaid responses not predicted by grating selectivity; and was consistent with broad anatomical selectivity observed in mouse V1. Our results show that visual feature binding can occur through local recurrent mechanisms without requiring feedforward convergence, and that such a mechanism is consistent with visual responses and cortical anatomy in mouse V1.Author summaryThe brain is a highly complex structure, with abundant connectivity between nearby neurons in the neocortex, the outermost and evolutionarily most recent part of the brain. Although the network architecture of the neocortex can appear disordered, connections between neurons seem to follow certain rules. These rules most likely determine how information flows through the neural circuits of the brain, but the relationship between particular connectivity rules and the function of the cortical network is not known. We built models of visual cortex in the mouse, assuming distinct rules for connectivity, and examined how the various rules changed the way the models responded to visual stimuli. We also recorded responses to visual stimuli of populations of neurons in anaesthetised mice, and compared these responses with our model predictions. We found that connections in neocortex probably follow a connectivity rule that groups together neurons that differ in simple visual properties, to build more complex representations of visual stimuli. This finding is surprising because primary visual cortex is assumed to support mainly simple visual representations. We show that including specific rules for non-random connectivity in cortical models, and precisely measuring those rules in cortical tissue, is essential to understanding how information is processed by the brain.


2004 ◽  
Vol 16 (6) ◽  
pp. 935-943 ◽  
Author(s):  
Krista Schendel ◽  
Lynn C. Robertson

Electrophysiological recordings in monkeys have now revealed several brain regions that contain bimodal visuotactile neurons capable of responding to either tactile or visual stimuli placed on or near the hands, arms, and face. These cells have now been found in frontal, parietal, and subcortical areas of the monkey brain, suggesting a cortical network of neurons that preferentially represent near peripersonal space. The degree to which the visual responses of such cells rely on input from the primary visual cortex and the extent to which they may contribute to visual perception is not completely understood. Nonetheless, recent neuropsychological studies suggest that a similar representation of near space may be bimodally coded in humans as well. Given the accumulating evidence for specialized processing of visual stimuli placed near the hands and arms, we hypothesized that arm position may be capable of modulating human visual ability. Here we report the case of WM, who lost his ability to see in his left visual hemifield after sustaining damage to his right primary visual cortex. Interestingly, the placement of WM's left arm into his “blind” field resulted in significantly better detection of left visual field stimuli compared to when his hand was placed in his lap at midline. Moreover, we found this attenuation to be confined to stimuli presented within reaching distance (unless a tool that extended WM's reach was held while he performed the test). These findings are highly consistent with the characteristics of the bimodal visuo-tactile neurons that have been described in monkeys. Thus, it seems that arm position can modulate human visual ability, even after damage to the primary visual cortex. This study provides an exciting bridge between monkey neurophysiology and human visual capacity while also offering a novel approach for improving visual defects acquired via cortical injury.


2017 ◽  
Vol 37 (13) ◽  
pp. 3532-3543 ◽  
Author(s):  
Megumi Kaneko ◽  
Yu Fu ◽  
Michael P. Stryker

2021 ◽  
Author(s):  
Megan A. Kirchgessner ◽  
Alexis D. Franklin ◽  
Edward M. Callaway

AbstractHigher-order (HO) thalamic nuclei interact extensively with the cerebral cortex and are innervated by excitatory corticothalamic (CT) populations in layers 5 and 6. While these distinct CT projections have long been thought to have different functional influences on the HO thalamus, this has never been directly tested. By optogenetically inactivating different CT populations in the primary visual cortex (V1) of awake mice, we demonstrate that layer 5, but not layer 6, CT projections drive visual responses in the HO visual pulvinar, even while both pathways provide retinotopic, baseline excitation to their thalamic targets. Inactivating the superior colliculus also suppressed visual responses in the pulvinar, demonstrating that cortical layer 5 and subcortical inputs both contribute to HO visual thalamic activity - even at the level of putative single neurons. Altogether, these results indicate a functional division of driver and modulator CT pathways from V1 to the visual thalamus in vivo.


Perception ◽  
1997 ◽  
Vol 26 (1_suppl) ◽  
pp. 106-106
Author(s):  
M Carandini ◽  
H B Barlow ◽  
A B Poirson ◽  
L P O'Keefe ◽  
J A Movshon

We tested the hypothesis that neurons in the primary visual cortex adapt selectively to contingencies in the attributes of visual stimuli. We recorded from single neurons in macaque V1 and measured the effects of adaptation either to the sum of two gratings (compound stimulus) or to the individual gratings. According to our hypothesis, there would be a component of adaptation that is specific to the compound stimulus. We performed two sets of experiments. In the first set one grating had optimal orientation and the other was orthogonal to it. In the second set the gratings were parallel, differed in spatial frequency, and were both effective in driving the cell. The first set of experiments, but not the second, provided evidence in favour of our hypothesis. In most cells tested with orthogonal gratings, adaptation to the compound stimulus reduced the responses to the compound stimulus more than the responses to the preferred grating. In addition, in most of these experiments the responses to the compound stimulus were reduced more by adaptation to the compound stimulus than by adaptation to the individual gratings. This suggests that a component of adaptation in the experiments with orthogonal gratings was specific to (and caused by) the contingent presence of the two gratings in the compound stimulus.


2021 ◽  
Author(s):  
Aaron M. Williams ◽  
Christopher F. Angeloni ◽  
Maria Neimark Geffen

In everyday life, we integrate visual and auditory information in routine tasks such as navigation and communication. While it is known that concurrent sound can improve visual perception, the neuronal correlates of this audiovisual integration are not fully understood. Specifically, it remains unknown whether improvement of the detection and discriminability of visual stimuli due to sound is reflected in the neuronal firing patterns in the primary visual cortex (V1). Furthermore, presentation of the sound can induce movement in the subject, but little is understood about whether and how sound-induced movement contributes to V1 neuronal activity. Here, we investigated how sound and movement interact to modulate V1 visual responses in awake, head-fixed mice and whether this interaction improves neuronal encoding of the visual stimulus. We presented visual drifting gratings with and without simultaneous auditory white noise to awake mice while recording mouse movement and V1 neuronal activity. Sound modulated the light-evoked activity of 80% of light-responsive neurons, with 95% of neurons exhibiting increased activity when the auditory stimulus was present. Sound consistently induced movement. However, a generalized linear model revealed that sound and movement had distinct and complementary effects of the neuronal visual responses. Furthermore, decoding of the visual stimulus from the neuronal activity was improved with sound, an effect that persisted even when controlling for movement. These results demonstrate that sound and movement modulate visual responses in complementary ways, resulting in improved neuronal representation of the visual stimulus. This study clarifies the role of movement as a potential confound in neuronal audiovisual responses, and expands our knowledge of how multi-modal processing is mediated at a neuronal level in the awake brain.


Sign in / Sign up

Export Citation Format

Share Document