scispace - formally typeset
Search or ask a question

Showing papers in "Journal of Vision in 2004"


Journal ArticleDOI
TL;DR: The results suggest that ordinary masking blocks feature detection, so the signal disappears, while crowding (like "illusory conjunction") is excessive feature integration - detected features are integrated over an inappropriately large area because there are no smaller integration fields - so the integrated signal is ambiguous, jumbled with the mask.
Abstract: A letter in the peripheral visual field is much harder to identify in the presence of nearby letters. This is “crowding.” Both crowding and ordinary masking are special cases of “masking,” which, in general, refers to any effect of a “mask” pattern on the discriminability of a signal. Here we characterize crowding, and propose a diagnostic test to distinguish it from ordinary masking. In ordinary masking, the signal disappears. In crowding, it remains visible, but is ambiguous, jumbled with its neighbors. Masks are usually effective only if they overlap the signal, but the crowding effect extends over a large region. The width of that region is proportional to signal eccentricity from the fovea and independent of signal size, mask size, mask contrast, signal and mask font, and number of masks. At 4 deg eccentricity, the threshold contrast for identification of a 0.32 deg signal letter is elevated (up to six-fold) by mask letters anywhere in a 2.3 deg region, 7 times wider than the signal. In ordinary masking, threshold contrast rises as a power function of mask contrast, with a shallow log-log slope of 0.5 to 1, whereas, in crowding, threshold is a sigmoidal function of mask contrast, with a steep log-log slope of 2 at close spacing. Most remarkably, although the threshold elevation decreases exponentially with spacing, the threshold and saturation contrasts of crowding are independent of spacing. Finally, ordinary masking is similar for detection and identification, but crowding occurs only for identification, not detection. More precisely, crowding occurs only in tasks that cannot be done based on a single detection by coarsely coded feature detectors. These results (and observers’ introspections) suggest that ordinary masking blocks feature detection, so the signal disappears, while crowding (like “illusory conjunction”) is excessive feature integration — detected features are integrated over an inappropriately large area because there are no smaller integration fields — so the integrated signal is ambiguous, jumbled with the mask. In illusory conjunction, observers see an object that is not there made up of features that are. A survey of the illusory conjunction literature finds that most of the illusory conjunction results are consistent with the spatial crowding described here, which depends on spatial proximity, independent of time pressure. The rest seem to arise through a distinct phenomenon that one might call “temporal crowding,” which depends on time pressure (“overloading attention”), independent of spatial proximity.

757 citations


Journal ArticleDOI
TL;DR: Findings from nine experiments in which VSTM for color, spatial frequency, and orientation was modeled using a signal detection theory (SDT) approach suggest that observers encode stimuli independently and in parallel, and that performance is limited by internal noise, which is a function of set size.
Abstract: Previous studies have suggested that visual short-term memory (VSTM) has a storage limit of approximately four items. However, the type of high-threshold (HT) model used to derive this estimate is based on a number of assumptions that have been criticized in other experimental paradigms (e.g., visual search). Here we report findings from nine experiments in which VSTM for color, spatial frequency, and orientation was modeled using a signal detection theory (SDT) approach. In Experiments 1-6, two arrays composed of multiple stimulus elements were presented for 100 ms with a 1500 ms ISI. Observers were asked to report in a yes/no fashion whether there was any difference between the first and second arrays, and to rate their confidence in their response on a 1-4 scale. In Experiments 1-3, only one stimulus element difference could occur (T = 1) while set size was varied. In Experiments 4-6, set size was fixed while the number of stimuli that might change was varied (T = 1, 2, 3, and 4). Three general models were tested against the receiver operating characteristics generated by the six experiments. In addition to the HT model, two SDT models were tried: one assuming summation of signals prior to a decision, the other using a max rule. In Experiments 7-9, observers were asked to directly report the relevant feature attribute of a stimulus presented 1500 ms previously, from an array of varying set size. Overall, the results suggest that observers encode stimuli independently and in parallel, and that performance is limited by internal noise, which is a function of set size.

710 citations


Journal ArticleDOI
TL;DR: It is concluded that objective methods of refraction based on wavefront aberration maps can accurately predict the results of subjective refraction and may be more precise and wavefront methods may become the new gold standard for specifying conventional and/or optimal corrections of refractive errors.
Abstract: We determined the accuracy and precision of 33 objective methods for predicting the results of conventional, sphero-cylindrical refraction from wavefront aberrations in a large population of 200 eyes. Accuracy for predicting defocus (as specified by the population mean error of prediction) varied from -0.50 D to +0.25 D across methods. Precision of these estimates (as specified by 95% limits of agreement) ranged from 0.5 to 1.0 D. All methods except one accurately predicted astigmatism to within +/-1/8D. Precision of astigmatism predictions was typically better than precision for predicting defocus and many methods were better than 0.5D. Paraxial curvature matching of the wavefront aberration map was the most accurate method for determining the spherical equivalent error whereas least-squares fitting of the wavefront was one of the least accurate methods. We argue that this result was obtained because curvature matching is a biased method that successfully predicts the biased endpoint stipulated by conventional refractions. Five methods emerged as reasonably accurate and among the most precise. Three of these were based on pupil plane metrics and two were based on image plane metrics. We argue that the accuracy of all methods might be improved by correcting for the systematic bias reported in this study. However, caution is advised because some tasks, including conventional refraction of defocus, require a biased metric whereas other tasks, such as refraction of astigmatism, are unbiased. We conclude that objective methods of refraction based on wavefront aberration maps can accurately predict the results of subjective refraction and may be more precise. If objective refractions are more precise than subjective refractions, then wavefront methods may become the new gold standard for specifying conventional and/or optimal corrections of refractive errors.

560 citations


Journal ArticleDOI
TL;DR: There is a growing empirical consensus that MLE provides a good quantitative account of cue combination and that sensory information is used in a manner that maximizes the precision of perceptual estimates.
Abstract: How does the visual system combine information from different depth cues to estimate 3D scene parameters? We tested a maximum-likelihood estimation (MLE) model of cue combination for perspective (texture) and binocular disparity cues to surface slant. By factoring the reliability of each cue into the combination process, MLE provides more reliable estimates of slant than would be available from either cue alone. We measured the reliability of each cue in isolation across a range of slants and distances using a slant -discrimination task. The reliability of the texture cue increases as |slant| increases and does not change with distance. The reliability of the disparity cue decreases as distance increases and varies with slant in a way that also depends on viewing distance. The trends in the single-cue data can be understood in terms of the information available in the retinal images and issues related to solving the binocular correspondence problem. To test the MLE model, we measured perceived slant of two-cue stimuli when disparity and texture were in conflict and the reliability of slant estimation when both cues were available. Results from the two-cue study indicate, consistent with the MLE model, that observers weight each cue according to its relative reliability: disparity weight decreased as distance and |slant| increased. We also observed the expected improvement in slant estimation when both cues were available. With few discrepancies, our data indicate that observers combine cues in a statistically optimal fashion and thereby reduce the variance of slant estimates below that which could be achieved from either cue alone. These results are consistent with other studies that quantitatively examined the MLE model of cue combination. Thus, there is a growing empirical consensus that MLE provides a good quantitative account of cue combination and that sensory information is used in a manner that maximizes the precision of perceptual estimates.

435 citations


Journal ArticleDOI
TL;DR: The results support the notion that binocular rivalry involves a more automatic, stimulus-driven form of visual competition than Necker cube reversal, and as a consequence, is less easily biased by selective attention.
Abstract: It is debated whether different forms of bistable perception result from common or separate neural mechanisms. Binocular rivalry involves perceptual alternations between competing monocular images, whereas ambiguous figures such as the Necker cube lead to alternations between two possible pictorial interpretations. Previous studies have shown that observers can voluntarily control the alternation rate of both rivalry and Necker cube reversal, perhaps suggesting that bistable perception results from a common mechanism of top-down selection. However, according to the biased competition model of selective attention, attention should be able to enhance the attended percept and suppress the unattended percept. Here, we investigated selective attentional modulation of dominance durations in bistable perception. Observers consistently showed much weaker selective attentional control for rivalry than for Necker cube reversal, even for rivalry displays that maximized the opportunities for feature-, object-, or space-based attentional selection. In contrast, nonselective control of alternation rate was comparably strong for both forms of bistable perception and corresponded poorly with estimates of selective attentional control. Our results support the notion that binocular rivalry involves a more automatic, stimulus-driven form of visual competition than Necker cube reversal, and as a consequence, is less easily biased by selective attention.

381 citations


Journal ArticleDOI
TL;DR: This work presents the correlation of 31 single-value metrics of optical quality to high-contrast visual acuity for 34 conditions where the RMS wavefront error was equal to 0.25 microm over a 6-mm pupil and finds the best metric, called the visual Strehl ratio, accounts for 81% of the variance in high- Contrast logMAR acuity.
Abstract: Wavefront-guided refractive surgery and custom optical corrections have reduced the residual root mean squared (RMS) wavefront error in the eye to relatively low levels (typically on the order of 0.25 microm or less over a 6-mm pupil, a dioptric equivalent of 0.19 D). It has been shown that experimental variation of the distribution of 0.25 microm of wavefront error across the pupil can cause variation in visual acuity of two lines on a standard logMAR acuity chart. This result demonstrates the need for single-value metrics other than RMS wavefront error to quantify the effects of low levels of aberration on acuity. In this work, we present the correlation of 31 single-value metrics of optical quality to high-contrast visual acuity for 34 conditions where the RMS wavefront error was equal to 0.25 microm over a 6-mm pupil. The best metric, called the visual Strehl ratio, accounts for 81% of the variance in high-contrast logMAR acuity.

361 citations


Journal ArticleDOI
TL;DR: It is shown that the dependency of specular distortions on the second derivative of the surface leads to distinctive fields of image orientation as the reflected world is warped across the surface, which are diagnostic of 3D shape and can be extracted from the image by populations of simple oriented filters.
Abstract: Many materials, including leaves, water, plastic, and chrome exhibit specular reflections It seems reasonable that the visual system can somehow exploit specular reflections to recover three-dimensional (3D) shape Previous studies (eg, J T Todd & E Mingolla, 1983; J F Norman, J T Todd, & G A Orban, 2004) have shown that specular reflections aid shape estimation, but the relevant image information has not yet been isolated Here we explain how specular reflections can provide reliable and accurate constraints on 3D shape We argue that the visual system can treat specularities somewhat like textures, by using the systematic patterns of distortion across the image of a specular surface to recover 3D shape However, there is a crucial difference between textures and specularities: In the case of textures, the image compressions depend on the first derivative of the surface depth (ie, surface orientation), whereas in the case of specularities, the image compressions depend on the second derivative (ie, surfaces curvatures) We suggest that this difference provides a cue that can help the visual system distinguish between textures and specularities, even when present simultaneously More importantly, we show that the dependency of specular distortions on the second derivative of the surface leads to distinctive fields of image orientation as the reflected world is warped across the surface We find that these "orientation fields" are (i) diagnostic of 3D shape, (ii) remain surprisingly stable when the world reflected in the surface is changed, and (iii) can be extracted from the image by populations of simple oriented filters Thus the use of specular reflections for 3D shape perception is both easier and more reliable than previous computational work would suggest

308 citations


Journal ArticleDOI
TL;DR: Adaptive optics testing supports the hypothesis that the neural visual system is adapted to the eye's aberrations, thereby removing somehow the effects of blur generated by the sensory apparatus from visual experience.
Abstract: A fundamental problem facing sensory systems is to recover useful information about the external world from signals that are corrupted by the sensory process itself. Retinal images in the human eye are affected by optical aberrations that cannot be corrected with ordinary spectacles or contact lenses, and the specific pattern of these aberrations is different in every eye. Though these aberrations always blur the retinal image, our subjective impression is that the visual world is sharp and clear, suggesting that the brain might compensate for their subjective influence. The recent introduction of adaptive optics to control the eye's aberrations now makes it possible to directly test this idea. If the brain compensates for the eye's aberrations, vision should be clearest with the eye's own aberrations rather than with unfamiliar ones. We asked subjects to view a stimulus through an adaptive optics system that either recreated their own aberrations or a rotated version of them. For all five subjects tested, the stimulus seen with the subject's own aberrations was always sharper than when seen through the rotated version. This supports the hypothesis that the neural visual system is adapted to the eye's aberrations, thereby removing somehow the effects of blur generated by the sensory apparatus from visual experience. This result could have important implications for methods to correct higher order aberrations with customized refractive surgery because some benefits of optimizing the correction optically might be undone by the nervous system's compensation for the old aberrations.

301 citations


Journal ArticleDOI
TL;DR: A typical eye will benefit over the entire accommodative range (0-6 D) if aberrations are corrected for distance viewing, even at high accommodative levels.
Abstract: Wave aberrations were measured with a Shack-Hartmann wavefront sensor (SHWS) in the right eye of a large young adult population when accommodative demands of 0, 3, and 6 D were presented to the tested eye through a Badal system. Three SHWS images were recorded at each accommodative demand and wave aberrations were computed over a 5-mm pupil (through 6th order Zernike polynomials). The accommodative response was calculated from the Zernike defocus over the central 3-mm diameter zone. Among all individual Zernike terms, spherical aberration showed the greatest change with accommodation. The change of spherical aberration was always negative, and was proportional to the change in accommodative response. Coma and astigmatism also changed with accommodation, but the direction of the change was variable. Despite the large inter-subject variability, the population average of the root mean square for all aberrations (excluding defocus) remained constant for accommodative levels up to 3.0 D. Even though aberrations change with accommodation, the magnitude of the aberration change remains less than the magnitude of the uncorrected aberrations, even at high accommodative levels. Therefore, a typical eye will benefit over the entire accommodative range (0-6 D) if aberrations are corrected for distance viewing.

292 citations


Journal ArticleDOI
TL;DR: Subjective judgment of best focus does not minimize RMS wavefront error, nor create paraxial focus, but makes the retina conjugate to a plane between these two, and it is possible to precisely predict subjective sphero-cylindrical refraction for monochromatic light using objective metrics.
Abstract: PURPOSE To determine the impact of higher-order monochromatic aberrations on lower-order subjective sphero-cylindrical refractions. METHODS Computationally-aberrated, monochromatic Sloan letters were presented on a high luminance display that was viewed by an observer through a 2.5mm pupil. Through-focus visual acuity (VA) was determined in the presence of spherical aberration (Z40) at three levels (0.10, 0.21 and 0.50D). Analogous through-astigmatism experiments measured visual acuity in the presence of secondary astigmatism (Z4+/-2) or coma (Z3-1). Measured visual acuity was correlated with 31 different metrics of image quality to determine which metric best predicts performance for degraded retinal images. The defocus and astigmatism levels that optimized each metric were compared with those that produced best visual acuity to determine which metric best predicts subjective refraction. RESULTS Spherical aberration, coma and secondary astigmatism all reduced VA and increased depth of focus. The levels of defocus and primary astigmatism that produced the best performance varied with levels of spherical aberration and secondary astigmatism, respectively. The presence of coma, however, did not affect cylindrical refraction. Image plane metrics, especially those that take into account the neural contrast sensitivity threshold (e.g. the visual Strehl ratio, VSOTF), are good predictors of visual acuity in both the through-focus and through-astigmatism experiments (R = -0.822 for VSOTF). Subjective sphero-cylindrical refractions were accurately predicted by some image-quality metrics (e.g., pupil fraction, VSOTF and standard deviation of PSF light distribution). CONCLUSION Subjective judgment of best focus does not minimize RMS wavefront error (Zernike defocus = 0), nor create paraxial focus (Seidel defocus = 0), but makes the retina conjugate to a plane between these two. It is possible to precisely predict subjective sphero-cylindrical refraction for monochromatic light using objective metrics.

288 citations


Journal ArticleDOI
TL;DR: Cross-sectional results do not support evidence of relationships between emmetropization and ocular aberrations, but are indicative of presbyopic changes occurring earlier in hyperopes than in myopes, and internal spherical aberration showed a significant age-related shift toward less negative values in the hyperopic group.
Abstract: This study investigated differences in geometrical properties and optical aberrations between a group of hyperopes and myopes (age-matched 30.3+/-5.2 and 30.5+/-3.8 years old, respectively, and with similar absolute refractive error 3.0+/-2.0 and -3.3+/-2.0, respectively). Axial length (AL) was measured by means of optical biometry, and corneal apical radius of curvature (CR) and asphericity (Q) were measured by fitting corneal topography data to biconic surfaces. Corneal aberrations were estimated from corneal topography by means of virtual ray tracing, and total aberrations were measured using a laser ray tracing technique. Internal aberrations were estimated by subtracting corneal from total aberrations. AL was significantly higher in myopes than in hyperopes and AL/CR was highly correlated with spherical equivalent. Hyperopic eyes tended to have higher (less negative) Q and higher total and corneal spherical aberration than myopic eyes. RMS for third-order aberrations was also significantly higher for the hyperopic eyes. Internal aberrations were not significantly different between the myopic and hyperopic groups, although internal spherical aberration showed a significant age-related shift toward less negative values in the hyperopic group. For these age and refraction ranges, our cross-sectional results do not support evidence of relationships between emmetropization and ocular aberrations. Our results may be indicative of presbyopic changes occurring earlier in hyperopes than in myopes.

Journal ArticleDOI
TL;DR: The results show that in normal indirect viewing, the impairment of character recognition by crowding is--in particular at intermediate flanker distances--caused to a large part by spatially imprecise focusing of attention, and lend psychophysical support to a separate coding of the what and where in pattern recognition.
Abstract: We studied mechanisms underlying the crowding effect in indirect form vision by measuring recognition contrast sensitivity of a character with flankers to the left and right. Attentional and featural contributions to the effect can be separated by a new paradigm that distinguishes pattern location errors from pattern recognition errors and further by manipulating the focusing of spatial attention through a positional cue, appearing 150 ms before the target. Measurements were on the horizontal meridian, at 1, 2, and 4 deg eccentricity, and a range of flankers' distances were used. Our results show that in normal indirect viewing, the impairment of character recognition by crowding is--in particular at intermediate flanker distances--caused to a large part by spatially imprecise focusing of attention. In contrast, the enhancement of performance by a transient positional cue seems mediated through a separate attentional mechanism such that attentional locus and focus are controlled independently. Our results furthermore lend psychophysical support to a separate coding of the what and where in pattern recognition.

Journal ArticleDOI
TL;DR: Strong evidence for compensation of horizontal/vertical (H/V) astigmatism (Zernike term Z5) lateral coma (Z8) and spherical aberration (Z12) is found, suggesting that it is actively determined by a fine-tuning process and not a passive result of genetically determined physiology.
Abstract: Both the anterior surface of the cornea and the internal optics (the posterior cornea, crystalline lens) contribute to the aberration of a wavefront passing through the eye. Artal, Guirao, Berrio, and Williams (2001) reported that the wavefront aberrations produced by the internal optics offset, or compensate for, the aberrations produced by the cornea to reduce ocular wavefront aberrations. We have investigated the wavefront aberrations of the cornea, internal optics, and complete eye on both the population and individual level to determine which aberrations are compensated and probable paths leading to that compensation. The corneal and ocular aberrations of 30 young subjects at relaxed accommodation were measured with the Topcon Wavefront Analyzer, which simultaneously measures refraction, corneal topography (videokeratoscope), and wavefront aberrations (Hartmann-Shack sensor). We found strong evidence for compensation of horizontal/vertical (H/V) astigmatism (Zernike term Z5) lateral coma (Z8) and spherical aberration (Z12). H/V astigmatism compensation is scaled for each individual, suggesting that it is actively determined by a fine-tuning process. Spherical aberration shows no individual compensation, suggesting that is a passive result of genetically determined physiology. Lateral coma shows individually scaled compensation, some of which may be attributable to eccentricity of the fovea.

Journal ArticleDOI
TL;DR: In summary, perceptual learning seems to rely at least partly on changes on a relatively early level of cortical information processing (early selection), such as the primary visual cortex under the influence of top-down influences (selection and shaping).
Abstract: Perceptual learning is any relatively permanent change of perception as a result of experience. Visual learning leads to sometimes dramatic and quite fast improvements of performance in perceptual tasks, such as hyperacuity discriminations. The improvement often is very specific for the exact task trained, for the precise stimulus orientation, the stimulus position in the visual field, and the eye used during training. This specificity indicates location of the underlying changes in the nervous system at least partly on the level of the primary visual cortex. The dependence of learning on error feedback and on attention, on the other hand, proves the importance of top-down influences from higher cortical centers. In summary, perceptual learning seems to rely at least partly on changes on a relatively early level of cortical information processing (early selection), such as the primary visual cortex under the influence of top-down influences (selection and shaping). An alternative explanation based on late selection is discussed.

Journal ArticleDOI
TL;DR: It is found that contrast learning is tuned to spatial frequency, orientation, retinal location, and, unexpectedly, contrast, and the contrast specificity of learning and the lack of learning under contrast roving provide new evidence in favor of a multiple contrast-selective channels model of contrast discrimination, and against saturating transducer models and multiplicative noise models.
Abstract: Unlike most visual tasks, contrast discrimination has been reported to be unchanged by practice (Dorais Sagi, 1997; Adini, Sagi, & Tsodyks, 2002), unless practice is undertaken in the presence of flankers (context-enabled learning, Adini et a[., 2002). Here we show that under experimental conditions nearly identical to those in the no-flanker practice experiment of Adini et al. (2002), practice significantly improved contrast discrimination. Moreover, in a separate experiment, we found that practice without flankers can improve contrast discrimination to a level only reached with flankers in Adini et al. (2002), but further practice with flankers produces no further improvement of contrast discrimination. These results call into question whether the "context-enabled learning" proposed by Adini et al. (2002) is different from regular contrast learning without flankers. In separate experiments, we found that contrast learning is tuned to spatial frequency, orientation, retinal location, and, unexpectedly, contrast. We also replicated Sagi, Adini, Tsodyks, and Wilkonsky's (2003) more recent finding that no regular contrast learning occurs if reference contrasts are randomly interleaved (contrast roving), and further demonstrated that flankers have no effect on contrast learning under contrast roving, another piece of evidence equating "context-enabled learning" to regular contrast learning. The contrast specificity of learning and the lack of learning under contrast roving provide new evidence in favor of a multiple contrast-selective channels model of contrast discrimination, and against saturating transducer models and multiplicative noise models.

Journal ArticleDOI
TL;DR: The maximum likelihood difference scaling procedure was used and it was found that gloss difference scales, when expressed in terms of the samples rather than the photometric characteristics, vary little with the direction of illumination.
Abstract: Gloss is an attribute of visual appearance that originates from the geometrical distribution of the light reflected by the surface. We used the maximum likelihood difference scaling (MLDS) procedure (L.T. Maloney & J. N. Yang, 2003) to estimate gloss scales over an extended range. Observers' judgments were obtained for a series of 10 black, coated samples for two directions of illumination, in binocular and monocular vision. The results showed a nonlinear relation between gloss percept and instrumental specular gloss values. Sensitivity is higher at extreme scale values than in the middle. In binocular vision, the sensitivity to gloss is higher than in monocular vision exclusively for high gloss levels. Lastly, we found that gloss difference scales, when expressed in terms of the samples rather than the photometric characteristics, vary little with the direction of illumination. Gloss scaling thus seems to be independent of the geometrical variations of the luminous flux at the surface of the sample. By analogy with the term "color constancy," we call this property "gloss constancy."

Journal ArticleDOI
TL;DR: The attentional cost associated with the visual discrimination of the gender of a face was investigated and results provide further evidence against the notion that only low-level representations can be accessed outside the focus of attention.
Abstract: The attentional cost associated with the visual discrimination of the gender of a face was investigated. Participants performed a face-gender discrimination task either alone (single-task) or concurrently (dual-task) with a known attentional demanding task (5-letter T/L discrimination). Overall performance on face-gender discrimination suffered remarkably little under the dual-task condition compared to the single-task condition. Similar results were obtained in experiments that controlled for potential training effects or the use of low-level cues in this discrimination task. Our results provide further evidence against the notion that only low-level representations can be accessed outside the focus of attention.

Journal ArticleDOI
TL;DR: It is suggested that encoding of orientation may have evolved to accommodate the anisotropy in natural scene content by perceptually discounting the most prevalent oriented content in a scene, thereby increasing the relative salience of objects and other content inA scene when viewed against a typical natural background.
Abstract: Many encoding mechanisms and processing strategies in the visual system appear to have evolved to better process the prevalent content in the visual world. Here we examine the relationship between the prevalence of natural scene content at different orientations and visual ability for detecting oriented natural scene content. Whereas testing with isolated gratings shows best performance at horizontal and vertical (the oblique effect), we report that when tested with natural scene content, performance is best at obliques and worst at horizontal (the horizontal effect). The present analysis of typical natural scenes shows that the prevalence of natural scene content matches the inverse of this horizontal effect pattern with most scene content at horizontal, next most at vertical, and least at obliques. We suggest that encoding of orientation may have evolved to accommodate the anisotropy in natural scene content by perceptually discounting the most prevalent oriented content in a scene, thereby increasing the relative salience of objects and other content in a scene when viewed against a typical natural background.

Journal ArticleDOI
TL;DR: This work illustrates, through analysis of photographically acquired, high dynamic range illumination maps, that real-world illumination possesses a high degree of statistical regularity, facilitating the understanding of human material perception, the design of robust computer vision systems, and the rendering of realistic computer graphics imagery.
Abstract: Although studies of vision and graphics often assume simple illumination models, real-world illumination is highly complex, with reflected light incident on a surface from almost every direction. One can capture the illumination from every direction at one point photographically using a spherical illumination map. This work illustrates, through analysis of photographically acquired, high dynamic range illumination maps, that real-world illumination possesses a high degree of statistical regularity. The marginal and joint wavelet coefficient distributions and harmonic spectra of illumination maps resemble those documented in the natural image statistics literature. However, illumination maps differ from typical photographs in that illumination maps are statistically nonstationary and may contain localized light sources that dominate their power spectra. Our work provides a foundation for statistical models of real-world illumination, thereby facilitating the understanding of human material perception, the design of robust computer vision systems, and the rendering of realistic computer graphics imagery.

Journal ArticleDOI
TL;DR: Whether processed as targets or non-targets, depending on the task, both human and animal face N170 were identical, and human faces in natural scenes elicit a clear but non-specific N170 that is not modulated by task status.
Abstract: The N170 is an event-related potential component reported to be very sensitive to human face stimuli. This study investigated the specificity of the N170, as well as its sensitivity to inversion and task status when subjects had to categorize either human or animal faces in the context of upright and inverted natural scenes. A conspicuous N170 was recorded for both face categories. Pictures of animal faces were associated with a N170 of similar amplitude compared to pictures of human faces, but with delayed peak latency. Picture inversion enhanced N170 amplitude for human faces and delayed its peak for both human and animal faces. Finally, whether processed as targets or non-targets, depending on the task, both human and animal face N170 were identical. Thus, human faces in natural scenes elicit a clear but non-specific N170 that is not modulated by task status. What appears to be specific to human faces is the strength of the inversion effect.

Journal ArticleDOI
TL;DR: It is concluded that the expression of visual implicit learning depends on attention but that latent learning of repeated information does not.
Abstract: Humans process a visual display more efficiently when they encounter it for a second time, showing learning of the display. This study tests whether implicit learning of complex visual contexts depends on attention. Subjects searched for a white target among black and white distractors. When the locations of the target and the attended set (white distractors) were repeated, search speed was enhanced, but when the locations of the target and the ignored set (black distractors) were repeated, search speed was unaffected. This suggests that the expression of learning depends on attention. However, during the transfer test, when the previously ignored set now was attended, it immediately facilitated performance. In contrast, when the previously attended set now was ignored, it no longer enhanced search speed. We conclude that the expression of visual implicit learning depends on attention but that latent learning of repeated information does not.

Journal ArticleDOI
TL;DR: This work investigated whether the mechanisms that mediate human color constancy embody this statistical regularity of the natural environment, so that constancy is best when the illuminant change is one likely to occur, and results indicate thatconstancy was least across the Red change, as one would expect for the statistics of natural daylight.
Abstract: The chromaticities of natural daylights cluster around the blackbody locus. We investigated whether the mechanisms that mediate human color constancy embody this statistical regularity of the natural environment, so that constancy is best when the illuminant change is one likely to occur. Observers viewed scenes displayed on a CRT-based stereoscope and adjusted a test patch embedded in the scene until it appeared achromatic. Scenes were rendered using physics-based graphics software (RADIANCE) coupled with custom extensions that ensured colorimetric accuracy. Across conditions, both the simulated illuminant and the simulated reflectance of scene objects were varied. Achromatic settings from paired conditions were used to compute a constancy index (CI) that characterizes the stability of object appearance across the two illuminants of the pair. Constancy indices were measured for four illuminant changes from a Neutral illuminant (CIE D65). Two of these changes (Blue and Yellow) were consistent with the statistics of daylight, whereas two (Green and Red) were not. The results indicate that constancy was least across the Red change, as one would expect for the statistics of natural daylight. Constancy for the Green direction, however, exceeded that for the Yellow illuminant change and was comparable to that for the Blue. This result is difficult to reconcile with the hypothesis that mechanisms of human constancy incorporate the statistics of daylights. Some possible reasons for the discrepancy are discussed.

Journal ArticleDOI
TL;DR: The failure of an LAM-based prediction - that the magnitude of learning-induced threshold reduction in high external noise must be less or equal to that in low external noise - is demonstrated.
Abstract: What is learned during perceptual learning? We address this question by analyzing how perceptual inefficiencies improve over the course of perceptual learning (Dosher & Lu, 1998). Systematic measurements of human performance as a function of both the amount of external noise added to the signal stimulus and the length of training received by the observers enable us to track changes of the characteristics of the perceptual system (e.g., internal noise[s] and efficiency of the perceptual template) as perceptual learning progresses, and, therefore, identifies the mechanism(s) underlying the observed performance improvements. Two different observer models, the linear amplifier model (LAM) and the perceptual template model (PTM), however, have led to two very different theories of learning mechanisms. Here we demonstrate the failure of an LAM-based prediction - that the magnitude of learning-induced threshold reduction in high external noise must be less or equal to that in low external noise. In Experiment 1, perceptual learning of Gabor orientation identification in fovea showed substantial performance improvements only in high external noise but not in zero or low noise. The LAM-based model was "forced" to account for the data with a combination of improved calculation efficiency and (paradoxical) compensatory increases of the equivalent internal noise. Based on the PTM framework, we conclude that perceptual learning in this task involved learning how to better exclude external noise, reflecting retuning of the perceptual template. The data provide the first empirical demonstration of an isolable mechanism of perceptual learning. This learning completely transferred to a different visual scale in a second experiment.

Journal ArticleDOI
TL;DR: The results confirm that above detection threshold, cross-channel masking involves contrast suppression and not (purely) mask-induced noise, and conclude that cross-Channel masking can be a powerful phenomenon, particularly at low test spatial frequencies and when mask and test are presented to different eyes.
Abstract: Masking is said to occur when a mask stimulus interferes with the visibility of a target (test) stimulus. One widely held view of this process supposes interactions between mask and test mechanisms (cross-channel masking), and explicit models (e.g., J. M. Foley, 1994) have proposed that the interactions are inhibitory. Unlike a within-channel model, where masking involves the combination of mask and test stimulus within a single mechanism, this cross-channel inhibitory model predicts that the mask should attenuate the perceived contrast of a test stimulus. Another possibility is that masking is due to an increase in noise, in which case, perception of contrast should be unaffected once the signal exceeds detection threshold. We use circular patches and annuli of sine-wave grating in contrast detection and contrast matching experiments to test these hypotheses and investigate interactions across spatial frequency, orientation, field position, and eye of origin. In both types of experiments we found substantial effects of masking that can occur over a factor of 3 in spatial frequency, 45° in orientation, across different field positions and between different eyes. We found the effects to be greatest at the lowest test spatial frequency we used (0.46 c/deg), and when the mask and test differed in all four dimensions simultaneously. This is surprising in light of previous work where it was concluded that suppression from the surround was strictly monocular (C. Chubb, G. Sperling, & J. A. Solomon, 1989). The results confirm that above detection threshold, cross-channel masking involves contrast suppression and not (purely) mask-induced noise. We conclude that cross-channel masking can be a powerful phenomenon, particularly at low test spatial frequencies and when mask and test are presented to different eyes. © 2004 ARVO.

Journal ArticleDOI
TL;DR: Six of the seven adult amblyopes repeatedly practiced a position discrimination task in the presence of positional noise and showed systematic and significant improvements in position acuity that were both eye and orientation specific.
Abstract: Adult amblyopes can improve positional acuity through practice; however, the neural mechanisms underlying this improvement are still not clear. In this study, seven adult amblyopes repeatedly practiced a position discrimination task in the presence of positional noise. We found that six of the seven showed systematic and significant improvements in position acuity that were both eye and orientation specific. Using a position-averaging model, we were able to parse the improvement in performance with practice into two factors: improvement in sampling efficiency and reduction of equivalent input noise. Three of the seven showed improved efficiency with no change in equivalent noise, two showed a significant reduction in equivalent noise with no change in efficiency, and one showed both improved efficiency and reduced equivalent noise. Interestingly, all observers showed substantial improvement in visual acuity, and one observer showed substantial improvement in stereoacuity. Three observers were also tested on a counting task, and all three improved after practicing positional discrimination. Our results reveal the mechanisms underlying perceptual learning in amblyopic vision, and may provide a basis for developing more effective and efficient strategies for the treatment of amblyopia.

Journal ArticleDOI
TL;DR: Evidence is provided that arousing sexual words involuntarily capture attention and enter awareness at the expense of goal-driven targets, at least in the context of laboratory experiments performed by young university participants for whom sexual material might have high impact and relevance.
Abstract: Participants are usually able to search rapid serial visual presentation (RSVP) streams and report a single target, given that RSVP distractors do not typically deplete attention required for target identification. Here, participants performed single target search, but the target was preceded by a to-be-ignored distractor varying in valence and arousal. When the critical distractor was a sexual word, lower target accuracy was observed, particularly at short distractor-target stimulus onset asynchronies, even when participants were shown the critical distractors beforehand and told to ignore them. No reduction in target accuracy was evidenced when the critical distractor was negative, positive, threatening, or emotionally neutral. Target accuracy was predicted by participants' arousal ratings to the critical distractor words and by their memory for them, but not by their valence ratings. Memory for critical distractors mediated the relationship between arousal and target accuracy. The results provide evidence that arousing sexual words involuntarily capture attention and enter awareness at the expense of goal-driven targets, at least in the context of laboratory experiments performed by young university participants for whom sexual material might have high impact and relevance.

Journal ArticleDOI
TL;DR: Tanaka et al. as mentioned in this paper found that Caucasian faces are processed more holistically than other-race faces, as a result of experience, while Asian subjects still presented a large other race effect, despite processing Caucasian faces as holistically as Asian faces.
Abstract: People are better at recognizing same-than other-race faces but the theoretical explanation of this phenomenon is still controversial. Here we tested the hypothesis that the "other-race effect'' is related to a reduced ability to encode configural information on other-race faces. Caucasian and Asian participants had to match whole faces to isolated facial features, or onto whole faces differing by a single feature, on both Caucasian and Asian faces. Participants performed better with whole faces as compared to isolated features, demonstrating a "holistic processing'' of faces (Tanaka & Farah, 1993). For Caucasian participants, this "whole/part advantage" was observed only for Caucasian faces. Asian participants who had been living for about a year among Caucasians had a comparable whole/part advantage regardless the race of the faces. These results indicate that same-race faces are processed more holistically than other-race faces, as a result of experience. However, despite processing Caucasian faces as holistically as Asian faces in this paradigm, Asian subjects still presented a large other-race effect. This observation suggests that holistic encoding may be a necessary step in order to be able to recognize other-race faces efficiently, but that it is by no means sufficient to overcome the other-race face effect.

Journal ArticleDOI
TL;DR: The model's robustness supports the hypothesis that the pendular nystagmus seen in CN is due to a loss of damping of the normal pursuit-system velocity oscillation (functionally, it is pursuit- system nyStagmus--PSN).
Abstract: Attempts to simulate dysfunction within ocular motor system (OMS) models capable of exhibiting known ocular motor behavior have provided valuable insight into the structure of the OMS required for normal visual function. The pendular waveforms of congenital nystagmus (CN) appear to be quite complex, composed of a sustained sinusoidal oscillation punctuated by braking saccades and foveating saccades followed by periods of extended foveation. Previously, we verified that these quick phases are generated by the same mechanism as voluntary saccades. We propose a computer model of the ocular motor system that simulates the responses of individuals with pendular CN (including its variable waveforms) based on the instability exhibited by the normal pursuit subsystem and its interaction with other components of the normal ocular motor control system. Fixation data from subjects with CN using both infrared and magnetic search coil oculography were used as templates for our simulations. Our OMS model simulates data from individuals with CN during fixation and in response to complex stimuli. The use of position and velocity efference copy to suppress oscillopsia is the key element in allowing for normal ocular motor behavior. The model's responses to target steps, pulse-steps, ramps, and step-ramps support the hypothetical explanation for the conditions that result in sustained pendular oscillation and the rules for the corrective saccadic responses that shape this underlying oscillation into the well-known family of pendular CN waveforms: pendular (P), pseudopendular (PP), pendular with foveating saccades (Pfs), and pseudopendular with foveating saccades (PPfs). Position error determined the saccadic amplitudes of foveating saccades, whereas stereotypical braking saccades were not dependent on visual information. Additionally, we propose a structure and method of operation for the fixation subsystem, and use it to prolong the low-velocity intervals immediately following foveating saccades. The model's robustness supports the hypothesis that the pendular nystagmus seen in CN is due to a loss of damping of the normal pursuit-system velocity oscillation (functionally, it is pursuit-system nystagmus--PSN).

Journal ArticleDOI
TL;DR: A very fast learning phase in contrast discrimination that improved performance under uncertainty is identified and it is contended that learned tasks that do not pass the uncertainty test involve modification of decision strategies that require exact knowledge of the stimulus.
Abstract: Performance in perceptual tasks improves with repetition (perceptual learning), eventually reaching a saturation level. Typically, when perceptual learning effects are studied, stimulus parameters are kept constant throughout the training and during the pre- and post-training tests. Here we investigate whether learning by repetition transfers to testing conditions in which the practiced stimuli are randomly interleaved during the post-training session. We studied practice effects with a contrast discrimination task, employing a number of training methods: (i) practice with a single, fixed pedestal (base-contrast), (ii) practice with several pedestals, and (iii) practice with several pedestals that included a spatial context. Pre- and post-training tests were carried out with the base contrast randomized across trials, under conditions of contrast uncertainty. The results showed that learning had taken place with the fixed pedestal method (i) and with the context method (iii), but only the latter survived the uncertainty test. In addition, we were able to identify a very fast learning phase in contrast discrimination that improved performance under uncertainty. We contend that learned tasks that do not pass the uncertainty test involve modification of decision strategies that require exact knowledge of the stimulus.