212 research outputs found
Parametric study of EEG sensitivity to phase noise during face processing
<b>Background: </b>
The present paper examines the visual processing speed of complex objects, here faces, by mapping the relationship between object physical properties and single-trial brain responses. Measuring visual processing speed is challenging because uncontrolled physical differences that co-vary with object categories might affect brain measurements, thus biasing our speed estimates. Recently, we demonstrated that early event-related potential (ERP) differences between faces and objects are preserved even when images differ only in phase information, and amplitude spectra are equated across image categories. Here, we use a parametric design to study how early ERP to faces are shaped by phase information. Subjects performed a two-alternative force choice discrimination between two faces (Experiment 1) or textures (two control experiments). All stimuli had the same amplitude spectrum and were presented at 11 phase noise levels, varying from 0% to 100% in 10% increments, using a linear phase interpolation technique. Single-trial ERP data from each subject were analysed using a multiple linear regression model.
<b>Results: </b>
Our results show that sensitivity to phase noise in faces emerges progressively in a short time window between the P1 and the N170 ERP visual components. The sensitivity to phase noise starts at about 120–130 ms after stimulus onset and continues for another 25–40 ms. This result was robust both within and across subjects. A control experiment using pink noise textures, which had the same second-order statistics as the faces used in Experiment 1, demonstrated that the sensitivity to phase noise observed for faces cannot be explained by the presence of global image structure alone. A second control experiment used wavelet textures that were matched to the face stimuli in terms of second- and higher-order image statistics. Results from this experiment suggest that higher-order statistics of faces are necessary but not sufficient to obtain the sensitivity to phase noise function observed in response to faces.
<b>Conclusion: </b>
Our results constitute the first quantitative assessment of the time course of phase information processing by the human visual brain. We interpret our results in a framework that focuses on image statistics and single-trial analyses
Luminance, colour, viewpoint and border enhanced disparity energy model
The visual cortex is able to extract disparity information through the use of binocular cells. This process is reflected by the Disparity Energy Model, which describes the role and functioning of simple and complex binocular neuron populations, and how they are able to extract disparity. This model uses explicit cell parameters to mathematically determine preferred cell disparities, like spatial frequencies, orientations, binocular phases and receptive field positions. However, the brain cannot access such explicit cell parameters; it must rely on cell responses. In this article, we implemented a trained binocular neuronal population, which encodes disparity information implicitly. This allows the population to learn how to decode disparities, in a similar way to how our visual system could have developed this ability during evolution. At the same time, responses of monocular simple and complex cells can also encode line and edge information, which is useful for refining disparities at object borders. The brain should then be able, starting from a low-level disparity draft, to integrate all information, including colour and viewpoint perspective, in order to propagate better estimates to higher cortical areas.Portuguese Foundation for Science and Technology (FCT); LARSyS FCT [UID/EEA/50009/2013]; EU project NeuroDynamics [FP7-ICT-2009-6, PN: 270247]; FCT project SparseCoding [EXPL/EEI-SII/1982/2013]; FCT PhD grant [SFRH-BD-44941-2008
Age-related delay in information accrual for faces: Evidence from a parametric, single-trial EEG approach
Background: In this study, we quantified age-related changes in the time-course of face processing
by means of an innovative single-trial ERP approach. Unlike analyses used in previous studies, our
approach does not rely on peak measurements and can provide a more sensitive measure of
processing delays. Young and old adults (mean ages 22 and 70 years) performed a non-speeded
discrimination task between two faces. The phase spectrum of these faces was manipulated
parametrically to create pictures that ranged between pure noise (0% phase information) and the
undistorted signal (100% phase information), with five intermediate steps.
Results: Behavioural 75% correct thresholds were on average lower, and maximum accuracy was
higher, in younger than older observers. ERPs from each subject were entered into a single-trial
general linear regression model to identify variations in neural activity statistically associated with
changes in image structure. The earliest age-related ERP differences occurred in the time window
of the N170. Older observers had a significantly stronger N170 in response to noise, but this age
difference decreased with increasing phase information. Overall, manipulating image phase
information had a greater effect on ERPs from younger observers, which was quantified using a
hierarchical modelling approach. Importantly, visual activity was modulated by the same stimulus
parameters in younger and older subjects. The fit of the model, indexed by R2, was computed at
multiple post-stimulus time points. The time-course of the R2 function showed a significantly slower
processing in older observers starting around 120 ms after stimulus onset. This age-related delay
increased over time to reach a maximum around 190 ms, at which latency younger observers had
around 50 ms time lead over older observers.
Conclusion: Using a component-free ERP analysis that provides a precise timing of the visual
system sensitivity to image structure, the current study demonstrates that older observers
accumulate face information more slowly than younger subjects. Additionally, the N170 appears to
be less face-sensitive in older observers
Chip-to-chip entanglement of transmon qubits using engineered measurement fields
© 2018 American Physical Society. While the on-chip processing power in circuit QED devices is growing rapidly, an open challenge is to establish high-fidelity quantum links between qubits on different chips. Here, we show entanglement between transmon qubits on different cQED chips with 49% concurrence and 73% Bell-state fidelity. We engineer a half-parity measurement by successively reflecting a coherent microwave field off two nearly identical transmon-resonator systems. By ensuring the measured output field does not distinguish |01) from |10), unentangled superposition states are probabilistically projected onto entangled states in the odd-parity subspace. We use in situ tunability and an additional weakly coupled driving field on the second resonator to overcome imperfect matching due to fabrication variations. To demonstrate the flexibility of this approach, we also produce an even-parity entangled state of similar quality, by engineering the matching of outputs for the |00) and |11) states. The protocol is characterized over a range of measurement strengths using quantum state tomography showing good agreement with a comprehensive theoretical model
Quantum Computing
Quantum mechanics---the theory describing the fundamental workings of
nature---is famously counterintuitive: it predicts that a particle can be in
two places at the same time, and that two remote particles can be inextricably
and instantaneously linked. These predictions have been the topic of intense
metaphysical debate ever since the theory's inception early last century.
However, supreme predictive power combined with direct experimental observation
of some of these unusual phenomena leave little doubt as to its fundamental
correctness. In fact, without quantum mechanics we could not explain the
workings of a laser, nor indeed how a fridge magnet operates. Over the last
several decades quantum information science has emerged to seek answers to the
question: can we gain some advantage by storing, transmitting and processing
information encoded in systems that exhibit these unique quantum properties?
Today it is understood that the answer is yes. Many research groups around the
world are working towards one of the most ambitious goals humankind has ever
embarked upon: a quantum computer that promises to exponentially improve
computational power for particular tasks. A number of physical systems,
spanning much of modern physics, are being developed for this task---ranging
from single particles of light to superconducting circuits---and it is not yet
clear which, if any, will ultimately prove successful. Here we describe the
latest developments for each of the leading approaches and explain what the
major challenges are for the future.Comment: 26 pages, 7 figures, 291 references. Early draft of Nature 464, 45-53
(4 March 2010). Published version is more up-to-date and has several
corrections, but is half the length with far fewer reference
Characterization of visual object representations in rat primary visual cortex
For most animal species, quick and reliable identification of visual objects is critical for survival. This applies also to rodents, which, in recent years, have become increasingly popular models of visual functions. For this reason in this work we analyzed how various properties of visual objects are represented in rat primary visual cortex (V1). The analysis has been carried out through supervised (classification) and unsupervised (clustering) learning methods. We assessed quantitatively the discrimination capabilities of V1 neurons by demonstrating how photometric properties (luminosity and object position in the scene) can be derived directly from the neuronal responses
The Invariance Hypothesis Implies Domain-Specific Regions in Visual Cortex
Is visual cortex made up of general-purpose information processing machinery, or does it consist of a collection of specialized modules? If prior knowledge, acquired from learning a set of objects is only transferable to new objects that share properties with the old, then the recognition system’s optimal organization must be one containing specialized modules for different object classes. Our analysis starts from a premise we call the invariance hypothesis: that the computational goal of the ventral stream is to compute an invariant-to-transformations and discriminative signature for recognition. The key condition enabling approximate transfer of invariance without sacrificing discriminability turns out to be that the learned and novel objects transform similarly. This implies that the optimal recognition system must contain subsystems trained only with data from similarly-transforming objects and suggests a novel interpretation of domain-specific regions like the fusiform face area (FFA). Furthermore, we can define an index of transformation-compatibility, computable from videos, that can be combined with information about the statistics of natural vision to yield predictions for which object categories ought to have domain-specific regions in agreement with the available data. The result is a unifying account linking the large literature on view-based recognition with the wealth of experimental evidence concerning domain-specific regions.National Science Foundation (U.S.). Science and Technology Center (Award CCF-1231216)National Science Foundation (U.S.) (Grant NSF-0640097)National Science Foundation (U.S.) (Grant NSF-0827427)United States. Air Force Office of Scientific Research (Grant FA8650-05-C-7262)Eugene McDermott Foundatio
Task and spatial frequency modulations of object processing: an EEG study.
Visual object processing may follow a coarse-to-fine sequence imposed by fast processing of low spatial frequencies (LSF) and slow processing of high spatial frequencies (HSF). Objects can be categorized at varying levels of specificity: the superordinate (e.g. animal), the basic (e.g. dog), or the subordinate (e.g. Border Collie). We tested whether superordinate and more specific categorization depend on different spatial frequency ranges, and whether any such dependencies might be revealed by or influence signals recorded using EEG. We used event-related potentials (ERPs) and time-frequency (TF) analysis to examine the time course of object processing while participants performed either a grammatical gender-classification task (which generally forces basic-level categorization) or a living/non-living judgement (superordinate categorization) on everyday, real-life objects. Objects were filtered to contain only HSF or LSF. We found a greater positivity and greater negativity for HSF than for LSF pictures in the P1 and N1 respectively, but no effects of task on either component. A later, fronto-central negativity (N350) was more negative in the gender-classification task than the superordinate categorization task, which may indicate that this component relates to semantic or syntactic processing. We found no significant effects of task or spatial frequency on evoked or total gamma band responses. Our results demonstrate early differences in processing of HSF and LSF content that were not modulated by categorization task, with later responses reflecting such higher-level cognitive factors
Recurrent Connections Aid Occluded Object Recognition by Discounting Occluders
Recurrent connections in the visual cortex are thought to aid object
recognition when part of the stimulus is occluded. Here we investigate if and
how recurrent connections in artificial neural networks similarly aid object
recognition. We systematically test and compare architectures comprised of
bottom-up (B), lateral (L) and top-down (T) connections. Performance is
evaluated on a novel stereoscopic occluded object recognition dataset. The task
consists of recognizing one target digit occluded by multiple occluder digits
in a pseudo-3D environment. We find that recurrent models perform significantly
better than their feedforward counterparts, which were matched in parametric
complexity. Furthermore, we analyze how the network's representation of the
stimuli evolves over time due to recurrent connections. We show that the
recurrent connections tend to move the network's representation of an occluded
digit towards its un-occluded version. Our results suggest that both the brain
and artificial neural networks can exploit recurrent connectivity to aid
occluded object recognition.Comment: 13 pages, 5 figures, accepted at the 28th International Conference on
Artificial Neural Networks, published in Springer Lecture Notes in Computer
Science vol 1172
A Preference for Contralateral Stimuli in Human Object- and Face-Selective Cortex
Visual input from the left and right visual fields is processed predominantly in the contralateral hemisphere. Here we investigated whether this preference for contralateral over ipsilateral stimuli is also found in high-level visual areas that are important for the recognition of objects and faces. Human subjects were scanned with functional magnetic resonance imaging (fMRI) while they viewed and attended faces, objects, scenes, and scrambled images in the left or right visual field. With our stimulation protocol, primary visual cortex responded only to contralateral stimuli. The contralateral preference was smaller in object- and face-selective regions, and it was smallest in the fusiform gyrus. Nevertheless, each region showed a significant preference for contralateral stimuli. These results indicate that sensitivity to stimulus position is present even in high-level ventral visual cortex
- …
