scispace - formally typeset
Search or ask a question

Showing papers on "Visual perception published in 2011"


Journal ArticleDOI
TL;DR: It is shown that visual search for a salient target is slowed by the presence of an inconspicuous, task-irrelevant item that was previously associated with monetary reward during a brief training session, suggesting that arbitrary and otherwise neutral stimuli imbued with value via associative learning capture attention powerfully and persistently during extinction, independently of goals and salience.
Abstract: Attention selects which aspects of sensory input are brought to awareness. To promote survival and well-being, attention prioritizes stimuli both voluntarily, according to context-specific goals (e.g., searching for car keys), and involuntarily, through attentional capture driven by physical salience (e.g., looking toward a sudden noise). Valuable stimuli strongly modulate voluntary attention allocation, but there is little evidence that high-value but contextually irrelevant stimuli capture attention as a consequence of reward learning. Here we show that visual search for a salient target is slowed by the presence of an inconspicuous, task-irrelevant item that was previously associated with monetary reward during a brief training session. Thus, arbitrary and otherwise neutral stimuli imbued with value via associative learning capture attention powerfully and persistently during extinction, independently of goals and salience. Vulnerability to such value-driven attentional capture covaries across individuals with working memory capacity and trait impulsivity. This unique form of attentional capture may provide a useful model for investigating failures of cognitive control in clinical syndromes in which value assigned to stimuli conflicts with behavioral goals (e.g., addiction, obesity).

877 citations


Journal ArticleDOI
TL;DR: It is reported that peripheral vision is limited with regard to pattern categorization by a distinctly lower representational complexity and processing speed than those imposed on low-level functions and by way of crowding.
Abstract: We summarize the various strands of research on peripheral vision and relate them to theories of form perception. After a historical overview, we describe quantifications of the cortical magnification hypothesis, including an extension of Schwartz's cortical mapping function. The merits of this concept are considered across a wide range of psychophysical tasks, followed by a discussion of its limitations and the need for non-spatial scaling. We also review the eccentricity dependence of other low-level functions including reaction time, temporal resolution, and spatial summation, as well as perimetric methods. A central topic is then the recognition of characters in peripheral vision, both at low and high levels of contrast, and the impact of surrounding contours known as crowding. We demonstrate how Bouma's law, specifying the critical distance for the onset of crowding, can be stated in terms of the retinocortical mapping. The recognition of more complex stimuli, like textures, faces, and scenes, reveals a substantial impact of mid-level vision and cognitive factors. We further consider eccentricity-dependent limitations of learning, both at the level of perceptual learning and pattern category learning. Generic limitations of extrafoveal vision are observed for the latter in categorization tasks involving multiple stimulus classes. Finally, models of peripheral form vision are discussed. We report that peripheral vision is limited with regard to pattern categorization by a distinctly lower representational complexity and processing speed. Taken together, the limitations of cognitive processing in peripheral vision appear to be as significant as those imposed on low-level functions and by way of crowding.

719 citations


Journal ArticleDOI
TL;DR: The goal of this review is to provide a broad-based synthesis of the most recent findings in crowding, to define what crowding is and is not, and to set the stage for future work that will extend the understanding of crowding well beyond low-level vision.

689 citations


Journal ArticleDOI
20 Oct 2011-Neuron
TL;DR: A high-dimensional model of the representational space in human ventral temporal (VT) cortex in which dimensions are response-tuning functions that are common across individuals and patterns of response are modeled as weighted sums of basis patterns associated with these response tunings is presented.

608 citations


Journal ArticleDOI
TL;DR: The overlay of significant meta-analyses resulted in extensive overlap in clusters, coupled with offset and unique clusters of reliable activity in areas of greatest overlap is the amygdala, followed by regions of medial prefrontal cortex, inferior frontal/orbitofrontal cortex, and extrastriate occipital cortex.

587 citations


Journal ArticleDOI
TL;DR: A population model for mid-ventral processing is developed, in which nonlinear combinations of V1 responses are averaged in receptive fields that grow with eccentricity, providing a quantitative framework for assessing the capabilities and limitations of everyday vision.
Abstract: Receptive fields of visual neurons get bigger along the ventral visual pathway and, in each area, they grow with distance from the fovea. The authors exploit these properties to build a model for visual representation in the ventral stream, using 'metameric' visual stimuli (which appear perceptually identical, but are actually different) to test the model predictions. The model can also explain deficits in peripheral recognition known as visual crowding.

576 citations


Journal ArticleDOI
TL;DR: In this article, the authors summarize salient developments in research on aging and vision over the past 25 years, focusing on spatial contrast sensitivity, vision under low luminance, temporal sensitivity and motion perception, and visual processing speed.

545 citations


Journal ArticleDOI
TL;DR: This work estimated observers' internal models for orientation and found that they matched the local orientation distribution measured in photographs, and determined how a neural population could embed probabilistic information responsible for such biases.
Abstract: Orientation judgments are more accurate at the horizontal and vertical orientations, possibly reflecting a statistical inference. Here the authors provide evidence for this idea, finding that observers' internal models for orientation match the local orientation distribution measured in photographs, and suggest how such information could be encoded in a neural population.

528 citations


Journal ArticleDOI
TL;DR: Results suggest that EEG recordings during normal vision are feasible and useful to consolidate findings from EEG and eye-tracking studies, and 4 technical and data-analytical problems that need to be addressed when FRPs are recorded in free-viewing situations are reviewed.
Abstract: Brain-electric correlates of reading have traditionally been studied with word-by-word presentation, a condition that eliminates important aspects of the normal reading process and precludes direct comparisons between neural activity and oculomotor behavior. In the present study, we investigated effects of word predictability on eye movements (EM) and fixation-related brain potentials (FRPs) during natural sentence reading. Electroencephalogram (EEG) and EM (via video-based eye tracking) were recorded simultaneously while subjects read heterogeneous German sentences, moving their eyes freely over the text. FRPs were time-locked to first-pass reading fixations and analyzed according to the cloze probability of the currently fixated word. We replicated robust effects of word predictability on EMs and the N400 component in FRPs. The data were then used to model the relation among fixation duration, gaze duration, and N400 amplitude, and to trace the time course of EEG effects relative to effects in EM behavior. In an extended Methodological Discussion section, we review 4 technical and data-analytical problems that need to be addressed when FRPs are recorded in free-viewing situations (such as reading, visual search, or scene perception) and propose solutions. Results suggest that EEG recordings during normal vision are feasible and useful to consolidate findings from EEG and eye-tracking studies.

441 citations


Journal ArticleDOI
09 Dec 2011-Science
TL;DR: It is suggested that early visual areas are so plastic that mere inductions of activity patterns are sufficient to cause visual perceptual learning (VPL).
Abstract: It is controversial whether the adult primate early visual cortex is sufficiently plastic to cause visual perceptual learning (VPL). The controversy occurs partially because most VPL studies have examined correlations between behavioral and neural activity changes rather than cause-and-effect relationships. With an online-feedback method that uses decoded functional magnetic resonance imaging (fMRI) signals, we induced activity patterns only in early visual cortex corresponding to an orientation without stimulus presentation or participants' awareness of what was to be learned. The induced activation caused VPL specific to the orientation. These results suggest that early visual areas are so plastic that mere inductions of activity patterns are sufficient to cause VPL. This technique can induce plasticity in a highly selective manner, potentially leading to powerful training and rehabilitative protocols.

425 citations


Journal ArticleDOI
TL;DR: This work investigates the influence of alpha activity on motion processing in two attentional conditions using magneto-encephalography and suggests that processing of inputs outside the field of attention is weakened by means of increased alpha activity.
Abstract: Because the human visual system is continually being bombarded with inputs, it is necessary to have effective mechanisms for filtering out irrelevant information. This is partly achieved by the allocation of attention, allowing the visual system to process relevant input while blocking out irrelevant input. What is the physiological substrate of attentional allocation? It has been proposed that alpha activity reflects functional inhibition. Here we asked if inhibition by alpha oscillations has behavioral consequences for suppressing the perception of unattended input. To this end, we investigated the influence of alpha activity on motion processing in two attentional conditions using magneto-encephalography. The visual stimuli used consisted of two random-dot kinematograms presented simultaneously to the left and right visual hemifields. Subjects were cued to covertly attend the left or right kinematogram. After 1.5 sec, a second cue tested whether subjects could report the direction of coherent motion in the attended (80%) or unattended hemifield (20%). Occipital alpha power was higher contralateral to the unattended side than to the attended side, thus suggesting inhibition of the unattended hemifield. Our key finding is that this alpha lateralization in the 20% invalidly cued trials did correlate with the perception of motion direction: Subjects with pronounced alpha lateralization were worse at detecting motion direction in the unattended hemifield. In contrast, lateralization did not correlate with visual discrimination in the attended visual hemifield. Our findings emphasize the suppressive nature of alpha oscillations and suggest that processing of inputs outside the field of attention is weakened by means of increased alpha activity.

Journal ArticleDOI
09 Jun 2011-Nature
TL;DR: By targeted genetic silencing of small subsets of cells in the Drosophila brain, it is shown that neurons in the ellipsoid body, but not in the mushroom bodies, are necessary for visual place learning.
Abstract: Insects such as ants or bees are renowned for their navigational prowess, which in part derives from their ability to learn and associate visual cues to locations in space. Now Charles Zuker and colleagues demonstrate that a powerful model organism — Drosophila melanogaster — is also capable of using vision to form spatial memories. By genetically silencing specific neurons, they then show that such spatial learning relies on a brain centre (the ellipsoid body) which is distinct from that used for non-spatial learning (the mushroom body). This work could lead to Drosophila becoming a model of choice for the study of spatial memory. The ability of insects to learn and navigate to specific locations in the environment has fascinated naturalists for decades. The impressive navigational abilities of ants, bees, wasps and other insects demonstrate that insects are capable of visual place learning1,2,3,4, but little is known about the underlying neural circuits that mediate these behaviours. Drosophila melanogaster (common fruit fly) is a powerful model organism for dissecting the neural circuitry underlying complex behaviours, from sensory perception to learning and memory. Drosophila can identify and remember visual features such as size, colour and contour orientation5,6. However, the extent to which they use vision to recall specific locations remains unclear. Here we describe a visual place learning platform and demonstrate that Drosophila are capable of forming and retaining visual place memories to guide selective navigation. By targeted genetic silencing of small subsets of cells in the Drosophila brain, we show that neurons in the ellipsoid body, but not in the mushroom bodies, are necessary for visual place learning. Together, these studies reveal distinct neuroanatomical substrates for spatial versus non-spatial learning, and establish Drosophila as a powerful model for the study of spatial memories.

Journal ArticleDOI
TL;DR: A review of the evidence concerning the occipital face area (OFA) can be found in this article, where it is shown to preferentially represent the parts of a face including the eyes, nose, and mouth.
Abstract: Functional magnetic resonance imaging (fMRI) studies have identified spatially distinct face-selective regions in human cortex. These regions have been linked together to form the components of a cortical network specialized for face perception but the cognitive operations performed in each region are not well understood. In this paper, we review the evidence concerning one of these face-selective regions, the occipital face area (OFA), to better understand what cognitive operations it performs in the face perception network. Neuropsychological evidence and transcranial magnetic stimulation (TMS) studies demonstrate the OFA is necessary for accurate face perception. fMRI and TMS studies investigating the functional role of the OFA suggest that it preferentially represents the parts of a face, including the eyes, nose, and mouth and that it does so at an early stage of visual perception. These studies are consistent with the hypothesis that the OFA is the first stage in a hierarchical face perception network in which the OFA represents facial components prior to subsequent processing of increasingly complex facial features in higher face-selective cortical regions.

Journal ArticleDOI
TL;DR: Electrocortical responses to neutral and emotional images during two phases suggest that distraction and reappraisal intervene at separate stages during emotion generation, a feature which may have distinct consequences that extend beyond the regulatory episode.

Journal ArticleDOI
TL;DR: Investigating the phase of prestimulus oscillatory activity found a systematic relationship between prepulse EEG phase and perceptual performance: phosphene probability changed by ∼15% between opposite phases, providing direct evidence for a chain of causal relations between thephase of ongoing oscillations, neuronal excitability, and visual perception.
Abstract: Why does neuronal activity in sensory brain areas sometimes give rise to perception, and sometimes not? Although neuronal noise is often invoked as the key factor, a portion of this variability could also be due to the history and current state of the brain affecting cortical excitability. Here we directly test this idea by examining whether the phase of prestimulus oscillatory activity is causally linked with modulations of cortical excitability and with visual perception. Transcranial magnetic stimulation (TMS) was applied over human visual cortex to induce illusory perceptions (phosphenes) while electroencephalograms (EEGs) were simultaneously recorded. Subjects reported the presence or absence of an induced phosphene following a single pulse of TMS at perceptual threshold. The phase of ongoing alpha (∼10 Hz) oscillations within 400 ms before the pulse significantly covaried with the perceptual outcome. This effect was observed in occipital regions around the site of TMS, as well as in a distant frontocentral region. In both regions, we found a systematic relationship between prepulse EEG phase and perceptual performance: phosphene probability changed by ∼15% between opposite phases. In summary, we provide direct evidence for a chain of causal relations between the phase of ongoing oscillations, neuronal excitability, and visual perception: ongoing oscillations create periodic “windows of excitability,” with sensory perception being more likely to occur at specific phases.

Journal ArticleDOI
TL;DR: It is argued that, like for many other aspects of vision, several different circuits related to salience, object recognition, actions, and value ultimately interact to determine gaze behavior.
Abstract: Eye movements are an integral and essential part of our human foveated vision system. Here, we review recent work on voluntary eye movements, with an emphasis on the last decade. More selectively, we address two of the most important questions about saccadic and smooth pursuit eye movements in natural vision. First, why do we saccade to where we do? We argue that, like for many other aspects of vision, several different circuits related to salience, object recognition, actions, and value ultimately interact to determine gaze behavior. Second, how are pursuit eye movements and perceptual experience of visual motion related? We show that motion perception and pursuit have a lot in common, but they also have separate noise sources that can lead to dissociations between them. We emphasize the point that pursuit actively modulates visual perception and that it can provide valuable information for motion perception.

Journal ArticleDOI
TL;DR: In a series of speeded classification tasks, spontaneous mappings between the auditory feature of pitch and the visual features of vertical location, size, and spatial frequency are found but not contrast.
Abstract: The brain may combine information from different sense modalities to enhance the speed and accuracy of detection of objects and events, and the choice of appropriate responses. There is mounting evidence that perceptual experiences that appear to be modality-specific are also influenced by activity from other sensory modalities, even in the absence of awareness of this interaction. In a series of speeded classification tasks, we found spontaneous mappings between the auditory feature of pitch and the visual features of vertical location, size, and spatial frequency but not contrast. By dissociating the task variables from the features that were cross-modally related, we find that the interactions happen in an automatic fashion and are possibly located at the perceptual level.

Journal ArticleDOI
25 May 2011-PLOS ONE
TL;DR: Full body illusions in which subjects experience the ownership of a doll's body and a giant's body are described and used as tools to demonstrate that the size of one's sensed own body directly influences the perception of object size and distance.
Abstract: A classical question in philosophy and psychology is if the sense of one's body influences how one visually perceives the world. Several theoreticians have suggested that our own body serves as a fundamental reference in visual perception of sizes and distances, although compelling experimental evidence for this hypothesis is lacking. In contrast, modern textbooks typically explain the perception of object size and distance by the combination of information from different visual cues. Here, we describe full body illusions in which subjects experience the ownership of a doll's body (80 cm or 30 cm) and a giant's body (400 cm) and use these as tools to demonstrate that the size of one's sensed own body directly influences the perception of object size and distance. These effects were quantified in ten separate experiments with complementary verbal, questionnaire, manual, walking, and physiological measures. When participants experienced the tiny body as their own, they perceived objects to be larger and farther away, and when they experienced the large-body illusion, they perceived objects to be smaller and nearer. Importantly, despite identical retinal input, this “body size effect” was greater when the participants experienced a sense of ownership of the artificial bodies compared to a control condition in which ownership was disrupted. These findings are fundamentally important as they suggest a causal relationship between the representations of body space and external space. Thus, our own body size affects how we perceive the world.

Journal ArticleDOI
TL;DR: The present article reviews recent findings on the role of prestimulus alpha oscillatory activity for visual perception and incorporates these results into a neurocognitive model that is able to account for various findings in temporal attention paradigms, specifically the attentional blink.

Journal ArticleDOI
TL;DR: This review examines some of the seminal work on the role of vision in the control of manual prehension and on the visual cues that play a critical role in this important human skill, and reviews the key evidence for the perception-action model.

Journal ArticleDOI
TL;DR: Recent advances in understanding the mouse visual system at the anatomical, receptive field and perceptual level are discussed, focusing on the opportunities and constraints those features provide toward the goal of understanding how vision works.

Journal ArticleDOI
TL;DR: These observations indicate that the earliest access to a high-level face representation does not precede the N170 onset in the human brain, and show that sensitivity to faces on the visual evoked potentials P1 and N1 (N170) is functionally dissociated.

Journal ArticleDOI
TL;DR: The nature and basis for human colour vision have been studied extensively, and ideas and principles that have been developed to account for norms and normalization in colour vision are drawn on to consider potential similarities and differences in the representation and adaptation of faces.
Abstract: The appearance of faces can be strongly affected by the characteristics of faces viewed previously. These perceptual after-effects reflect processes of sensory adaptation that are found throughout the visual system, but which have been considered only relatively recently in the context of higher level perceptual judgements. In this review, we explore the consequences of adaptation for human face perception, and the implications of adaptation for understanding the neural-coding schemes underlying the visual representation of faces. The properties of face after-effects suggest that they, in part, reflect response changes at high and possibly face-specific levels of visual processing. Yet, the form of the after-effects and the norm-based codes that they point to show many parallels with the adaptations and functional organization that are thought to underlie the encoding of perceptual attributes like colour. The nature and basis for human colour vision have been studied extensively, and we draw on ideas and principles that have been developed to account for norms and normalization in colour vision to consider potential similarities and differences in the representation and adaptation of faces.

Journal ArticleDOI
TL;DR: This work has revealed that the pattern of aftereffects is similar across many stimulus dimensions, pointing to common coding principles in which adaptation plays a central role.
Abstract: Visual coding is a highly dynamic process and continuously adapting to the current viewing context. The perceptual changes that result from adaptation to recently viewed stimuli remain a powerful and popular tool for analyzing sensory mechanisms and plasticity. Over the last decade, the footprints of this adaptation have been tracked to both higher and lower levels of the visual pathway and over a wider range of timescales, revealing that visual processing is much more adaptable than previously thought. This work has also revealed that the pattern of aftereffects is similar across many stimulus dimensions, pointing to common coding principles in which adaptation plays a central role. However, why visual coding adapts has yet to be fully answered.

Journal ArticleDOI
TL;DR: This work derived a simple parametric decoder assuming neuronal independence and a more sophisticated empirical decoder that learned the structure of the measured neuronal response distributions, including their correlated variability, indicating that their structure contains critical information for sensory decoding.
Abstract: Visual function depends on the accuracy of signals carried by visual cortical neurons. Combining information across neurons should improve this accuracy because single neuron activity is variable. We examined the reliability of information inferred from populations of simultaneously recorded neurons in macaque primary visual cortex. We considered a decoding framework that computes the likelihood of visual stimuli from a pattern of population activity by linearly combining neuronal responses and tested this framework for orientation estimation and discrimination. We derived a simple parametric decoder assuming neuronal independence and a more sophisticated empirical decoder that learned the structure of the measured neuronal response distributions, including their correlated variability. The empirical decoder used the structure of these response distributions to perform better than its parametric variant, indicating that their structure contains critical information for sensory decoding. These results show how neuronal responses can best be used to inform perceptual decision-making.

Journal ArticleDOI
TL;DR: Together, these results suggest visual capacity limitation is due to discrete, slot-like, resources, each containing limited pools of neural information that can be divided among objects.
Abstract: Cognition has a severely limited capacity: Adult humans can retain only about four items “in mind”. This limitation is fundamental to human brain function: Individual capacity is highly correlated with intelligence measures and capacity is reduced in neuropsychiatric diseases. Although human capacity limitations are well studied, their mechanisms have not been investigated at the single-neuron level. Simultaneous recordings from monkey parietal and frontal cortex revealed that visual capacity limitations occurred immediately upon stimulus encoding and in a bottom-up manner. Capacity limitations were found to reflect a dual model of working memory. The left and right halves of visual space had independent capacities and thus are discrete resources. However, within each hemifield, neural information about successfully remembered objects was reduced by adding further objects, indicating that resources are shared. Together, these results suggest visual capacity limitation is due to discrete, slot-like, resources, each containing limited pools of neural information that can be divided among objects.

Journal ArticleDOI
TL;DR: These results identify a set of candidate frontal, parietal and subcortical regions that integrate visual and tactile information for the multisensory perception of one's own hand.
Abstract: In the non-human primate brain, a number of multisensory areas have been described where individual neurons respond to visual, tactile and bimodal visuotactile stimulation of the upper limb. It has been shown that such bimodal neurons can integrate sensory inputs in a linear or nonlinear fashion. In humans, activity in a similar set of brain regions has been associated with visuotactile stimulation of the hand. However, little is known about how these areas integrate visual and tactile information. In this functional magnetic resonance imaging experiment, we employed tactile, visual, and visuotactile stimulation of the right hand in an ecologically valid setup where participants were looking directly at their upper limb. We identified brain regions that were activated by both visual and tactile stimuli as well as areas exhibiting greater activity in the visuotactile condition than in both unisensory ones. The posterior and inferior parietal, dorsal, and ventral premotor cortices, as well as the cerebellum, all showed evidence of multisensory linear (additive) responses. Nonlinear, superadditive responses were observed in the cortex lining the left anterior intraparietal sulcus, the insula, dorsal premotor cortex, and, subcortically, the putamen. These results identify a set of candidate frontal, parietal and subcortical regions that integrate visual and tactile information for the multisensory perception of one's own hand.

Journal ArticleDOI
TL;DR: Using functional magnetic resonance imaging, it is shown that activation during Braille reading in blind individuals peaks in the VWFA, with striking anatomical consistency within and between blind and sighted.

Journal ArticleDOI
TL;DR: The basic concepts and logic of ideal observer analysis are summarized and applications in a number of different areas are described, including pattern detection, discrimination and estimation, perceptual grouping, shape, depth and motion perception and visual attention.

Journal ArticleDOI
TL;DR: It is demonstrated that long-term music training is related to improvements in working memory, in both the auditory and visual domains and in terms of both behavioral and ERP measures.