scispace - formally typeset
Search or ask a question

Showing papers on "Visual perception published in 2013"


Journal ArticleDOI
TL;DR: The results of the review support the conclusion that aesthetic response can be studied rigorously and meaningfully within the framework of scientific psychology.
Abstract: Human aesthetic preference in the visual domain is reviewed from definitional, methodological, empirical, and theoretical perspectives. Aesthetic science is distinguished from the perception of art and from philosophical treatments of aesthetics. The strengths and weaknesses of important behavioral techniques are presented and discussed, including two-alternative forced-choice, rank order, subjective rating, production/adjustment, indirect, and other tasks. Major findings are reviewed about preferences for colors (single colors, color combinations, and color harmony), spatial structure (low-level spatial properties, shape properties, and spatial composition within a frame), and individual differences in both color and spatial structure. Major theoretical accounts of aesthetic response are outlined and evaluated, including explanations in terms of mere exposure effects, arousal dynamics, categorical prototypes, ecological factors, perceptual and conceptual fluency, and the interaction of multiple components. The results of the review support the conclusion that aesthetic response can be studied rigorously and meaningfully within the framework of scientific psychology.

451 citations


Journal ArticleDOI
TL;DR: This work explores the behavior of mice under a visual display that simulates an approaching object, which causes defensive reactions in some other species, and shows that mice respond to this stimulus either by initiating escape within a second or by freezing for an extended period.

442 citations


Journal ArticleDOI
03 May 2013-Science
TL;DR: A neural decoding approach is presented in which machine-learning models predict the contents of visual imagery during the sleep-onset period by discovering links between human functional magnetic resonance imaging patterns and verbal reports with the assistance of lexical and image databases.
Abstract: Visual imagery during sleep has long been a topic of persistent speculation, but its private nature has hampered objective analysis. Here we present a neural decoding approach in which machine-learning models predict the contents of visual imagery during the sleep-onset period, given measured brain activity, by discovering links between human functional magnetic resonance imaging patterns and verbal reports with the assistance of lexical and image databases. Decoding models trained on stimulus-induced brain activity in visual cortical areas showed accurate classification, detection, and identification of contents. Our findings demonstrate that specific visual experience during sleep is represented by brain activity patterns shared by stimulus perception, providing a means to uncover subjective contents of dreaming using objective neural measurement.

417 citations



Journal ArticleDOI
TL;DR: This article found that many voxels across occipito-temporal and fronto-parietal cortex shifted their tuning toward the attended category during visual search for different object categories in natural movies.
Abstract: Little is known about how attention changes the cortical representation of sensory information in humans. On the basis of neurophysiological evidence, we hypothesized that attention causes tuning changes to expand the representation of attended stimuli at the cost of unattended stimuli. To investigate this issue, we used functional magnetic resonance imaging to measure how semantic representation changed during visual search for different object categories in natural movies. We found that many voxels across occipito-temporal and fronto-parietal cortex shifted their tuning toward the attended category. These tuning shifts expanded the representation of the attended category and of semantically related, but unattended, categories, and compressed the representation of categories that were semantically dissimilar to the target. Attentional warping of semantic representation occurred even when the attended category was not present in the movie; thus, the effect was not a target-detection artifact. These results suggest that attention dynamically alters visual representation to optimize processing of behaviorally relevant objects during natural vision.

328 citations


Journal ArticleDOI
TL;DR: The composite face effect as discussed by the authors has been used in more than 60 studies that have provided information about the specificity and nature of perceptual integration between facial parts, the impairment of this process in acquired prosopagnosia, its developmental course, temporal dynamics, and neural basis.
Abstract: Two identical top halves of a face are perceived as being different when their bottom halves belong to different faces, showing that the parts of a face cannot be perceived independently from the whole face. When this visual illusion is inserted in a matching task, observers make more mistakes and/or are slower at matching identical top face halves aligned with different bottom halves than when the bottom halves are spatially offset: The composite face effect. This composite face paradigm has been used in more than 60 studies that have provided information about the specificity and nature of perceptual integration between facial parts (“holistic face perception”), the impairment of this process in acquired prosopagnosia, its developmental course, temporal dynamics, and neural basis. Following a review of the main contributions made with the paradigm, I explain its rationale and strengths, and discuss its methodological parameters, making a number of proposals for its optimal use and refinement in order to improve our understanding of holistic face perception. Finally, I explain how this standard composite face paradigm is fundamentally different than the application to facial parts of a congruency/interference paradigm that has a long tradition in experimental psychology since Stroop (1935), and which was originally developed to measure attentional and response interference between different representations rather than perceptual integration. Moreover, a version of this congruency/interference paradigm used extensively over the past years with composite faces lacks a baseline measure and has decisional, attentional, and stimulus confounds, making the findings of these studies impossible to interpret in terms of holistic perception. I conclude by encouraging researchers in this field to concentrate fully on the standard composite face paradigm, gaze contingency, and other behavioural measures that can help us take one of the most important challenges of visual perception research: Understanding the neural mechanisms of holistic face perception.

320 citations


Journal ArticleDOI
TL;DR: The results demonstrate that the contents of VSTM can be decoded from areas that exhibit a transient response to visual stimuli, but not from regions that exhibit elevated, sustained load-sensitive delay-period activity, and indicate load sensitivity of the patterns of activity that support V STM maintenance.
Abstract: Traditionally, load sensitivity of sustained, elevated activity has been taken as an index of storage for a limited number of items in visual short-term memory (VSTM). Recently, studies have demonstrated that the contents of a single item held in VSTM can be decoded from early visual cortex, despite the fact that these areas do not exhibit elevated, sustained activity. It is unknown, however, whether the patterns of neural activity decoded from sensory cortex change as a function of load, as one would expect from a region storing multiple representations. Here, we use multivoxel pattern analysis to examine the neural representations of VSTM in humans across multiple memory loads. In an important extension of previous findings, our results demonstrate that the contents of VSTM can be decoded from areas that exhibit a transient response to visual stimuli, but not from regions that exhibit elevated, sustained load-sensitive delay-period activity. Moreover, the neural information present in these transiently activated areas decreases significantly with increasing load, indicating load sensitivity of the patterns of activity that support VSTM maintenance. Importantly, the decrease in classification performance as a function of load is correlated with within-subject changes in mnemonic resolution. These findings indicate that distributed patterns of neural activity in putatively sensory visual cortex support the representation and precision of information in VSTM.

311 citations


Journal ArticleDOI
TL;DR: A synthetic stimuli replicating the higher-order statistical dependencies found in natural texture images was constructed and used to stimulate macaque V1 and V2 neurons, revealing a particular functional role for V2 in the representation of natural image structure.
Abstract: There is no generally accepted account of the function of the second visual cortical area (V2), partly because no simple response properties robustly distinguish V2 neurons from those in primary visual cortex (V1). We constructed synthetic stimuli replicating the higher-order statistical dependencies found in natural texture images, and used them to stimulate macaque V1 and V2 neurons. Most V2 cells responded more vigorously to these textures than to control stimuli lacking naturalistic structure; V1 cells did not. fMRI measurements in humans revealed differences between V1 and V2 that paralleled the neuronal measurements. The ability of human observers to detect naturalistic structure in different types of texture was well predicted by the strength of neuronal and fMRI responses in V2 but not in V1. Together, these results reveal a novel and particular role for V2 in the representation of natural image structure.

310 citations


Journal ArticleDOI
TL;DR: Results show a strong correspondence between psychophysical and neurophysiological data, suggesting that the phase entrainment of low-frequency oscillations to external sensory cues can serve as an important and flexible mechanism for enhancing sensory processing.
Abstract: Although it is increasingly accepted that temporal expectation can modulate early perceptual processing, the underlying neural computations remain unknown. In the present study, we combined a psychophysical paradigm with electrophysiological recordings to investigate the putative contribution of low-frequency oscillatory activity in mediating the modulation of visual perception by temporal expectation. Human participants judged the orientation of brief targets (visual Gabor patterns tilted clockwise or counterclockwise) embedded within temporally regular or irregular streams of noise-patches used as temporal cues. Psychophysical results indicated that temporal expectation enhanced the contrast sensitivity of visual targets. A diffusion model indicated that rhythmic temporal expectation modulated the signal-to-noise gain of visual processing. The concurrent electrophysiological data revealed that the phase of delta oscillations overlying human visual cortex (1–4 Hz) was predictive of the quality of target processing only in regular streams of events. Moreover, in the regular condition, the optimum phase of these perception-predictive oscillations occurred in anticipation of the expected events. Together, these results show a strong correspondence between psychophysical and neurophysiological data, suggesting that the phase entrainment of low-frequency oscillations to external sensory cues can serve as an important and flexible mechanism for enhancing sensory processing.

297 citations


Journal ArticleDOI
TL;DR: It is found that gray matter volumes of the frontal polar and precuneus regions themselves correlated across individuals, and a formal model comparison analysis suggested that this structural covariation was sufficient to account for the behavioral correlation of metacognition in the two tasks.
Abstract: A recent study found that, across individuals, gray matter volume in the frontal polar region was correlated with visual metacognition capacity (i.e., how well one's confidence ratings distinguish between correct and incorrect judgments). A question arises as to whether the putative metacognitive mechanisms in this region are also used in other metacognitive tasks involving, for example, memory. A novel psychophysical measure allowed us to assess metacognitive efficiency separately in a visual and a memory task, while taking variations in basic task performance capacity into account. We found that, across individuals, metacognitive efficiencies positively correlated between the two tasks. However, voxel-based morphometry analysis revealed distinct brain structures for the two kinds of metacognition. Replicating a previous finding, variation in visual metacognitive efficiency was correlated with volume of frontal polar regions. However, variation in memory metacognitive efficiency was correlated with volume of the precuneus. There was also a weak correlation between visual metacognitive efficiency and precuneus volume, which may account for the behavioral correlation between visual and memory metacognition (i.e., the precuneus may contain common mechanisms for both types of metacognition). However, we also found that gray matter volumes of the frontal polar and precuneus regions themselves correlated across individuals, and a formal model comparison analysis suggested that this structural covariation was sufficient to account for the behavioral correlation of metacognition in the two tasks. These results highlight the importance of the precuneus in higher-order memory processing and suggest that there may be functionally distinct metacognitive systems in the human brain.

241 citations


Journal ArticleDOI
TL;DR: A systematic, large-scale structure is found in the neural responses related to the interaction between two major cognitive dimensions of object representation: animacy and real-world size that reflect the major joints in the representational structure of objects and place informative constraints on the nature of the underlying cognitive architecture.
Abstract: Occipito-temporal cortex is known to house visual object representations, but the organization of the neural activation patterns along this cortex is still being discovered. Here we found a systematic, large-scale structure in the neural responses related to the interaction between two major cognitive dimensions of object representation: animacy and real-world size. Neural responses were measured with functional magnetic resonance imaging while human observers viewed images of big and small animals and big and small objects. We found that real-world size drives differential responses only in the object domain, not the animate domain, yielding a tripartite distinction in the space of object representation. Specifically, cortical zones with distinct response preferences for big objects, all animals, and small objects, are arranged in a spoked organization around the occipital pole, along a single ventromedial, to lateral, to dorsomedial axis. The preference zones are duplicated on the ventral and lateral surface of the brain. Such a duplication indicates that a yet unknown higher-order division of labor separates object processing into two substreams of the ventral visual pathway. Broadly, we suggest that these large-scale neural divisions reflect the major joints in the representational structure of objects and thus place informative constraints on the nature of the underlying cognitive architecture.

Journal ArticleDOI
TL;DR: This review discusses recent progress in identifying the specific prefrontal circuits controlling visual attention and its neural correlates within the primate visual system and examines the persisting challenge of precisely defining how behavior should be affected when attentional function is lost.
Abstract: The faculty of attention endows us with the capacity to process important sensory information selectively while disregarding information that is potentially distracting. Much of our understanding of the neural circuitry underlying this fundamental cognitive function comes from neurophysiological studies within the visual modality. Past evidence suggests that a principal function of the prefrontal cortex (PFC) is selective attention and that this function involves the modulation of sensory signals within posterior cortices. In this review, we discuss recent progress in identifying the specific prefrontal circuits controlling visual attention and its neural correlates within the primate visual system. In addition, we examine the persisting challenge of precisely defining how behavior should be affected when attentional function is lost.

Journal ArticleDOI
TL;DR: The findings suggest that facilitated detection of invisible objects due to language occurs at a perceptual rather than semantic locus, and when information associated with verbal labels matches stimulus-driven activity, language can provide a boost to perception, propelling an otherwise invisible image into awareness.
Abstract: Linguistic labels (e.g., “chair”) seem to activate visual properties of the objects to which they refer. Here we investigated whether language-based activation of visual representations can affect the ability to simply detect the presence of an object. We used continuous flash suppression to suppress visual awareness of familiar objects while they were continuously presented to one eye. Participants made simple detection decisions, indicating whether they saw any image. Hearing a verbal label before the simple detection task changed performance relative to an uninformative cue baseline. Valid labels improved performance relative to no-label baseline trials. Invalid labels decreased performance. Labels affected both sensitivity (d′) and response times. In addition, we found that the effectiveness of labels varied predictably as a function of the match between the shape of the stimulus and the shape denoted by the label. Together, the findings suggest that facilitated detection of invisible objects due to language occurs at a perceptual rather than semantic locus. We hypothesize that when information associated with verbal labels matches stimulus-driven activity, language can provide a boost to perception, propelling an otherwise invisible image into awareness.

Journal ArticleDOI
TL;DR: It is posited that the beta-band activity decrease recorded in low performing elderly subjects reflects difficulty in activation and deficits in sustaining attentional processes.

BookDOI
01 Feb 2013
TL;DR: This article presented new discoveries about infants' visual perception in areas ranging from sensory processes to visual cognition at the 23rd Carnegie Mellon Symposia on Cognition (CMCS) in 2003.
Abstract: The chapters in this book are based on papers presented at the 23rd Carnegie Mellon Symposia on Cognition. At this exciting event, speaker after speaker presented new discoveries about infants' visual perception in areas ranging from sensory processes to visual cognition. The field continues to make significant progress in understanding the infant's perceptual world. Several advances have come from the development of new methods for exploring infant perception and cognition that have brought new empirical findings. Advances have also been made in understanding the mechanisms underlying perceptual development. Outstanding examples of this ongoing progress can be seen in the chapters of this volume.

Journal ArticleDOI
TL;DR: 3-D models from motion captured actions acted out by two people, e.g., hugging, are created and the degree to which velocities of joints are able to predict the recognition performance of social interactions is examined in order to determine candidate visual cues underlying the recognition of social interaction recognition.
Abstract: Recognizing social interactions, e.g. two people shaking hands, is important for obtaining information about other people and the surrounding social environment. Despite the visual complexity of social interactions, humans have often little difficulties to visually recognize social interactions. What is the visual representation of social interactions and the bodily visual cues that promote this remarkable human ability? Viewpoint dependent representations are considered to be at the heart of the visual recognition of many visual stimuli including objects (Bulthoff & Edelman, 1992), and biological motion patterns (Verfaillie, 1993). Here we addressed the question whether complex social actions acted out between pairs of people, e.g. hugging, are also represented in a similar manner. To this end, we created 3-D models from motion captured actions acted out by two people, e.g. hugging. These 3-D models allowed to present the same action from different viewpoints. Participants task was to discriminate a target action from distractor actions using a one-interval-forced-choice (1IFC) task. We measured participants' recognition performance in terms of reaction times (RT) and d-prime (d'). For each tested action we found one view that lead to superior recognition performance compared to other views. This finding demonstrates view-dependent effects of visual recognition, which are in line with the idea of a view dependent representations of social interactions. Subsequently, we examined the degree to which velocities of joints are able to predict the recognition performance of social interactions in order to determine candidate visual cues underlying the recognition of social interactions. We found that the velocities of the right arm, lower left leg, and both feet correlated with recognition performance.

Journal ArticleDOI
TL;DR: The results suggest that the well-established effect of contour on aesthetic preference can be extended to architecture, and the combination of behavioral and neural evidence underscores the role of emotion in the authors' preference for curvilinear objects in this domain.
Abstract: On average, we urban dwellers spend about 90% of our time indoors, and share the intuition that the physical features of the places we live and work in influence how we feel and act. However, there is surprisingly little research on how architecture impacts behavior, much less on how it influences brain function. To begin closing this gap, we conducted a functional magnetic resonance imaging study to examine how systematic variation in contour impacts aesthetic judgments and approach-avoidance decisions, outcome measures of interest to both architects and users of spaces alike. As predicted, participants were more likely to judge spaces as beautiful if they were curvilinear than rectilinear. Neuroanatomically, when contemplating beauty, curvilinear contour activated the anterior cingulate cortex exclusively, a region strongly responsive to the reward properties and emotional salience of objects. Complementing this finding, pleasantness—the valence dimension of the affect circumplex—accounted for nearly 60% of the variance in beauty ratings. Furthermore, activation in a distributed brain network known to underlie the aesthetic evaluation of different types of visual stimuli covaried with beauty ratings. In contrast, contour did not affect approach-avoidance decisions, although curvilinear spaces activated the visual cortex. The results suggest that the well-established effect of contour on aesthetic preference can be extended to architecture. Furthermore, the combination of our behavioral and neural evidence underscores the role of emotion in our preference for curvilinear objects in this domain.

Journal ArticleDOI
26 Jul 2013
TL;DR: The principles and methods of modern algorithms for automatically predicting the quality of visual signals are discussed and divided into understandable modeling subproblems by casting the problem as analogous to assessing the efficacy of a visual communication system.
Abstract: Finding ways to monitor and control the perceptual quality of digital visual media has become a pressing concern as the volume being transported and viewed continues to increase exponentially. This paper discusses the principles and methods of modern algorithms for automatically predicting the quality of visual signals. By casting the problem as analogous to assessing the efficacy of a visual communication system, it is possible to divide the quality assessment problem into understandable modeling subproblems. Along the way, we will visit models of natural images and videos, of visual perception, and a broad spectrum of applications.

Journal ArticleDOI
TL;DR: The role of prestimulus power fluctuations in two visual phenomena called double-flash illusion (DFI) and fusion effect (FE), both consisting of suprathreshold stimuli, is studied to suggest that reduced alpha-band power does not always predict improved visual processing, but rather enhanced excitability.
Abstract: Several studies have demonstrated that prestimulus occipital alpha-band activity substantially influences subjective perception and discrimination of near-threshold or masked visual stimuli. Here, we studied the role of prestimulus power fluctuations in two visual phenomena called double-flash illusion (DFI) and fusion effect (FE), both consisting of suprathreshold stimuli. In both phenomena, human subjects' perception varies on a trial-by-trial basis between perceiving one or two visual stimuli, despite constant stimulation. In the FE, two stimuli correspond to veridical perception. In the DFI, two stimuli correspond to an illusory perception. This provides for a critical test of whether reduced alpha power indeed promotes veridical perception in general. We find that in both, DFI and FE, reduced prestimulus occipital alpha predicts the perception of two stimuli, regardless of whether this is veridical (FE) or illusory (DFI). Our results suggest that reduced alpha-band power does not always predict improved visual processing, but rather enhanced excitability. In addition, for the DFI, enhanced prestimulus occipital gamma-band power predicted the perception of two visual stimuli. These findings provide new insights into the role of prestimulus rhythmic activity for visual processing.

Journal ArticleDOI
TL;DR: It is found that suppressing activity in V1 substantially impaired performance in visual detection tasks and depended on the retinotopic position of the visual stimulus, confirming that the effect was due to the specific suppression of the visually driven V1 neurons.
Abstract: In mammals, the lateral geniculate nucleus (LGN) and the superior colliculus (SC) are the major targets of visual inputs from the retina. The LGN projects mainly to primary visual cortex (V1) while the SC targets the thalamus and brainstem, providing two potential pathways for processing visual inputs. Indeed, cortical lesion experiments in rodents have yielded mixed results, leading to the hypothesis that performance of simple visual behaviors may involve computations performed entirely by this subcortical pathway through the SC. However, these previous experiments have been limited by both their assays of behavioral performance and their use of lesions to change cortical activity. To determine the contribution of V1 to these tasks, we trained mice to perform threshold detection tasks in which they reported changes in either the contrast or orientation of visual stimuli. We then reversibly inhibited V1 by optogenetically activating parvalbumin-expressing inhibitory neurons with channelrhodopsin-2. We found that suppressing activity in V1 substantially impaired performance in visual detection tasks. The behavioral deficit depended on the retinotopic position of the visual stimulus, confirming that the effect was due to the specific suppression of the visually driven V1 neurons. Behavioral effects were seen with only moderate changes in neuronal activity, as inactivation that raised neuronal contrast thresholds by a median of only 14% was associated with a doubling of behavioral contrast detection threshold. Thus, detection of changes in either orientation or contrast is dependent on, and highly sensitive to, the activity of neurons in V1.

Journal ArticleDOI
29 Mar 2013-PLOS ONE
TL;DR: In this article, the alpha-frequency (10.6 Hz) was found to have specific behavioral consequences, relative to stimulation at control frequencies (3.9 Hz, 7.1 Hz, 14.2 Hz), leading to alpha-band oscillations in visual performance measures that correlate in precise frequency across individuals with resting alpha-rhythms recorded over parieto-occipital areas.
Abstract: Oscillations are an important aspect of neuronal activity. Interestingly, oscillatory patterns are also observed in behaviour, such as in visual performance measures after the presentation of a brief sensory event in the visual or another modality. These oscillations in visual performance cycle at the typical frequencies of brain rhythms, suggesting that perception may be closely linked to brain oscillations. We here investigated this link for a prominent rhythm of the visual system (the alpha-rhythm, 8–12 Hz) by applying rhythmic visual stimulation at alpha-frequency (10.6 Hz), known to lead to a resonance response in visual areas, and testing its effects on subsequent visual target discrimination. Our data show that rhythmic visual stimulation at 10.6 Hz: 1) has specific behavioral consequences, relative to stimulation at control frequencies (3.9 Hz, 7.1 Hz, 14.2 Hz), and 2) leads to alpha-band oscillations in visual performance measures, that 3) correlate in precise frequency across individuals with resting alpha-rhythms recorded over parieto-occipital areas. The most parsimonious explanation for these three findings is entrainment (phase-locking) of ongoing perceptually relevant alpha-band brain oscillations by rhythmic sensory events. These findings are in line with occipital alpha-oscillations underlying periodicity in visual performance, and suggest that rhythmic stimulation at frequencies of intrinsic brain-rhythms can be used to reveal influences of these rhythms on task performance to study their functional roles.

Journal ArticleDOI
TL;DR: A new model of high-level visual cortex consisting of ventral, lateral, and dorsal components, where multimodal processing related to vision, action, haptics, and language converges in the lateral pathway is concluded.
Abstract: Neurophysiology and optical imaging studies in monkeys and functional magnetic resonance imaging (fMRI) studies in both monkeys and humans have localized clustered neural responses in inferotemporal cortex selective for images of biologically relevant categories, such as faces and limbs. Using higher resolution (1.5 mm voxels) fMRI scanning methods than past studies (3–5 mm voxels), we recently reported a network of multiple face- and limb-selective regions that neighbor one another in human ventral temporal cortex (Weiner and Grill-Spector, Neuroimage, 52(4):1559–1573, 2010) and lateral occipitotemporal cortex (Weiner and Grill-Spector, Neuroimage, 56(4):2183–2199, 2011). Here, we expand on three basic organization principles of high-level visual cortex revealed by these findings: (1) consistency in the anatomical location of functional regions, (2) preserved spatial relationship among functional regions, and (3) a topographic organization of face- and limb-selective regions in adjacent and alternating clusters. We highlight the implications of this structure in comparing functional brain organization between typical and atypical populations. We conclude with a new model of high-level visual cortex consisting of ventral, lateral, and dorsal components, where multimodal processing related to vision, action, haptics, and language converges in the lateral pathway. Electronic supplementary material The online version of this article (doi:10.1007/s00426-011-0392-x) contains supplementary material, which is available to authorized users.

Journal ArticleDOI
TL;DR: Experimental results show that compared with the relevant existing metrics, the proposed metric can achieve higher consistency with subjective assessment of stereoscopic images.
Abstract: Perceptual quality assessment is a challenging issue in 3D signal processing research. It is important to study 3D signal directly instead of studying simple extension of the 2D metrics directly to the 3D case as in some previous studies. In this paper, we propose a new perceptual full-reference quality assessment metric of stereoscopic images by considering the binocular visual characteristics. The major technical contribution of this paper is that the binocular perception and combination properties are considered in quality assessment. To be more specific, we first perform left-right consistency checks and compare matching error between the corresponding pixels in binocular disparity calculation, and classify the stereoscopic images into non-corresponding, binocular fusion, and binocular suppression regions. Also, local phase and local amplitude maps are extracted from the original and distorted stereoscopic images as features in quality assessment. Then, each region is evaluated independently by considering its binocular perception property, and all evaluation results are integrated into an overall score. Besides, a binocular just noticeable difference model is used to reflect the visual sensitivity for the binocular fusion and suppression regions. Experimental results show that compared with the relevant existing metrics, the proposed metric can achieve higher consistency with subjective assessment of stereoscopic images.

Journal ArticleDOI
TL;DR: High-density electrophysiology from a cohort of high-functioning children with ASD was recorded while they performed a simple AV reaction time task, and evidence for processing differences was seen as early as 100 ms poststimulation, and topographic analysis suggested thatChildren with ASD relied on different cortical networks during this early multisensory processing stage.
Abstract: Successful integration of auditory and visual inputs is crucial for both basic perceptual functions and for higher-order processes related to social cognition. Autism spectrum disorders (ASD) are characterized by impairments in social cognition and are associated with abnormalities in sensory and perceptual processes. Several groups have reported that individuals with ASD are impaired in their ability to integrate socially relevant audiovisual (AV) information, and it has been suggested that this contributes to the higher-order social and cognitive deficits observed in ASD. However, successful integration of auditory and visual inputs also influences detection and perception of nonsocial stimuli, and integration deficits may impair earlier stages of information processing, with cascading downstream effects. To assess the integrity of basic AV integration, we recorded high-density electrophysiology from a cohort of high-functioning children with ASD (7-16 years) while they performed a simple AV reaction time task. Children with ASD showed considerably less behavioral facilitation to multisensory inputs, deficits that were paralleled by less effective neural integration. Evidence for processing differences relative to typically developing children was seen as early as 100 ms poststimulation, and topographic analysis suggested that children with ASD relied on different cortical networks during this early multisensory processing stage.

Journal ArticleDOI
TL;DR: It is shown that GABA, the main inhibitory neurotransmitter, shapes the dynamics of bistable perception, paving the way for future studies into the competitive neural interactions across the visual cortical hierarchy that elicit conscious perception.

Journal ArticleDOI
TL;DR: In this article, the authors examined the relationship between abundant in-store stimuli and limited human perceptual capacity and found that consumers have fragmented visual attention during grocery shopping, and that their visual attention is simultaneously influenced and disrupted by the shelf display.

BookDOI
15 Apr 2013
TL;DR: In this article, the authors present a model of the structure of visual objects in terms of the perception of shape and orientation, which is based on the Gestalts and their components: Nature of Information-Precedence.
Abstract: Contents: J.L.F. Trespalacios, Object Perception and Phenomenology. H.E. Egeth, C.L. Folk, P.A. Mullin, Spatial Parallelism in the Processing of Lines, Letters, and Lexicality. J.R. Pomerantz, E.A. Pristach,C.E. Carson, Attention and Object Perception. L.A. Cooper, Mental Models of the Structure of Visual Objects. S.E. Palmer, Reference Frames in the Perception of Shape and Orientation. M.I. Lasaga, Gestalts and Their Components: Nature of Information-Precedence. B.E. Shepp, On Perceiving Objects: Holistic Versus Featural Properties. S. Ballesteros, Some Determinants of Perceived Structure: Effects of Stimulus and Tasks. G.R. Lockhead, Category Bounds and Stimulus Variability. J.D. Smith, Analytic and Holistic Processes in Categorization. L.B. Smith, P. Evans, Similarity, Identity, and Dimensions: Perceptual Classification in Children and Adults. D.G.K. Nelson, The Nature and Occurrence of Holistic Processing. T.B. Ward, Analytic and Holistic Modes of Processing in Category Learning.

Journal ArticleDOI
TL;DR: The findings demonstrate that people actually depend primarily on visual information when making judgments about music performance, and the dominance of visual information emerges to the degree that it is overweighted relative to auditory information, even when sound is consciously valued as the core domain content.
Abstract: Social judgments are made on the basis of both visual and auditory information, with consequential implications for our decisions. To examine the impact of visual information on expert judgment and its predictive validity for performance outcomes, this set of seven experiments in the domain of music offers a conservative test of the relative influence of vision versus audition. People consistently report that sound is the most important source of information in evaluating performance in music. However, the findings demonstrate that people actually depend primarily on visual information when making judgments about music performance. People reliably select the actual winners of live music competitions based on silent video recordings, but neither musical novices nor professional musicians were able to identify the winners based on sound recordings or recordings with both video and sound. The results highlight our natural, automatic, and nonconscious dependence on visual cues. The dominance of visual information emerges to the degree that it is overweighted relative to auditory information, even when sound is consciously valued as the core domain content.

Journal ArticleDOI
TL;DR: This work found neural correlates of confidence in the primate pulvinar, a visual thalamic nucleus that has been expanding markedly through evolution that encode a subject's certainty of visual categorization and contribute to perceptual confidence.
Abstract: When we recognize a sensory event, we experience a confident feeling that we certainly know the perceived world 'here and now'. However, it is unknown how and where the brain generates such 'perceptual confidence'. Here we found neural correlates of confidence in the primate pulvinar, a visual thalamic nucleus that has been expanding markedly through evolution. During a categorization task, the majority of pulvinar responses did not correlate with any 'perceptual content'. During an opt-out task, pulvinar responses decreased when monkeys chose 'escape' options, suggesting less confidence in their perceptual categorization. Functional silencing of the pulvinar increased monkeys' escape choices in the opt-out task without affecting categorization performance; this effect was specific to the contralateral visual target. These data were supported by a theoretical model of confidence, indicating that pulvinar activities encode a subject's certainty of visual categorization and contribute to perceptual confidence.

Journal ArticleDOI
29 May 2013-PLOS ONE
TL;DR: It is demonstrated that it is possible to classify the task that a person is engaged in from their eye movements using multivariate pattern classification, which has important theoretical implications for computational and neural models of eye movement control.
Abstract: In human vision, acuity and color sensitivity are greatest at the center of fixation and fall off rapidly as visual eccentricity increases. Humans exploit the high resolution of central vision by actively moving their eyes three to four times each second. Here we demonstrate that it is possible to classify the task that a person is engaged in from their eye movements using multivariate pattern classification. The results have important theoretical implications for computational and neural models of eye movement control. They also have important practical implications for using passively recorded eye movements to infer the cognitive state of a viewer, information that can be used as input for intelligent human-computer interfaces and related applications.