scholarly journals Modeling the Influence of Visual Priming in Feature and Conjunctive Search

2017 ◽  
Vol 17 (10) ◽  
pp. 1133
Author(s):  
Jordan Haggit ◽  
Joseph Houpt
2000 ◽  
Author(s):  
Trent Welsh ◽  
Anita Sarno ◽  
Gabriele Gratton ◽  
Monica Fabiani
Keyword(s):  

PLoS ONE ◽  
2015 ◽  
Vol 10 (12) ◽  
pp. e0144730 ◽  
Author(s):  
Xiong Jiang ◽  
Yang Jiang ◽  
Raja Parasuraman
Keyword(s):  

PLoS ONE ◽  
2010 ◽  
Vol 5 (11) ◽  
pp. e14087 ◽  
Author(s):  
Clayton Hickey ◽  
Leonardo Chelazzi ◽  
Jan Theeuwes

2021 ◽  
Author(s):  
Giulio Matteucci ◽  
Benedetta Zattera ◽  
Rosilari Bellacosa Marotti ◽  
Davide Zoccolan

AbstractComputing global motion direction of extended visual objects is a hallmark of primate high-level vision. Although neurons selective for global motion have also been found in mouse visual cortex, it remains unknown whether rodents can combine multiple motion signals into global, integrated percepts. To address this question, we trained two groups of rats to discriminate either gratings (G group) or plaids (i.e., superpositions of gratings with different orientations; P group) drifting horizontally along opposite directions. After the animals learned the task, we applied a visual priming paradigm, where presentation of the target stimulus was preceded by the brief presentation of either a grating or a plaid. The extent to which rat responses to the targets were biased by such prime stimuli provided a measure of the spontaneous, perceived similarity between primes and targets. We found that gratings and plaids, when uses as primes, were equally effective at biasing the perception of plaid direction for the rats of the P group. Conversely, for G group, only the gratings acted as effective prime stimuli, while the plaids failed to alter the perception of grating direction. To interpret these observations, we simulated a decision neuron reading out the representations of gratings and plaids, as conveyed by populations of either component or pattern cells (i.e., local or global motion detectors). We concluded that the findings for the P group are highly consistent with the existence of a population of pattern cells, playing a functional role similar to that demonstrated in primates. We also explored different scenarios that could explain the failure of the plaid stimuli to elicit a sizable priming magnitude for the G group. These simulations yielded testable predictions about the properties of motion representations in rodent visual cortex at the single-cell and circuitry level, thus paving the way to future neurophysiology experiments.


2021 ◽  
Author(s):  
Elena Pyatigorskaya ◽  
Matteo Maran ◽  
Emiliano Zaccarella

Language comprehension proceeds at a very fast pace. It is argued that context influences the speed of language comprehension by providing informative cues for the correct processing of the incoming linguistic input. Priming studies investigating the role of context in language processing have shown that humans quickly recognise target words that share orthographic, morphological, or semantic information with their preceding primes. How syntactic information influences the processing of incoming words is however less known. Early syntactic priming studies reported faster recognition for noun and verb targets (e.g., apple or sing) following primes with which they form grammatical phrases or sentences (the apple, he sings). The studies however leave open a number of questions about the reported effect, including the degree of automaticity of syntactic priming, the facilitative versus inhibitory nature, and the specific mechanism underlying the priming effect—that is, the type of syntactic information primed on the target word. Here we employed a masked syntactic priming paradigm in four behavioural experiments in German language to test whether masked primes automatically facilitate the categorization of nouns and verbs presented as flashing visual words. Overall, we found robust syntactic priming effects with masked primes—thus suggesting high automaticity of the process—but only when verbs were morpho-syntactically marked (er kau-t; he chew-s). Furthermore, we found that, compared to baseline, primes slow down target categorisation when the relationship between prime and target is syntactically incorrect, rather than speeding it up when the prime-target relationship is syntactically correct. This argues in favour of an inhibitory nature of syntactic priming. Overall, the data indicate that humans automatically extract abstract syntactic features from word categories as flashing visual words, which has an impact on the speed of successful language processing during language comprehension.


2022 ◽  
pp. 107-131
Author(s):  
Dhruti P. Sharma ◽  
Devesh C. Jinwala

E-health is a cloud-based system to store and share medical data with the stakeholders. From a security perspective, the stored data are in encrypted form that could further be searched by the stakeholders through searchable encryption (SE). Practically, an e-health system with support of multiple stakeholders (that may work as either data owner [writer] or user [reader]) along with the provision of multi-keyword search is desirable. However, the existing SE schemes either support multi-keyword search in multi-reader setting or offer multi-writer, multi-reader mechanism along with single-keyword search only. This chapter proposes a multi-keyword SE for an e-health system in multi-writer multi-reader setting. With this scheme, any registered writer could share data with any registered reader with optimal storage-computational overhead on writer. The proposed scheme offers conjunctive search with optimal search complexity at server. It also ensures security to medical records and privacy of keywords. The theoretical and empirical analysis demonstrates the effectiveness of the proposed work.


2016 ◽  
Vol 62 ◽  
pp. 242-255 ◽  
Author(s):  
Joel Harman ◽  
Ross Brown ◽  
Daniel Johnson ◽  
Stefanie Rinderle-Ma ◽  
Udo Kannengiesser

2016 ◽  
Vol 124 (2) ◽  
pp. 549-563 ◽  
Author(s):  
Sumi Shigeno

This study examined the effects of preceding contextual stimuli, either auditory or visual, on the identification of spoken target words. Fifty-one participants (29% males, 71% females; mean age = 24.5 years, SD = 8.5) were divided into three groups: no context, auditory context, and visual context. All target stimuli were spoken words masked with white noise. The relationships between the context and target stimuli were as follows: identical word, similar word, and unrelated word. Participants presented with context experienced a sequence of six context stimuli in the form of either spoken words or photographs. Auditory and visual context conditions produced similar results, but the auditory context aided word identification more than the visual context in the similar word relationship. We discuss these results in the light of top-down processing, motor theory, and the phonological system of language.


Sign in / Sign up

Export Citation Format

Share Document