scispace - formally typeset
Search or ask a question
Topic

Human visual system model

About: Human visual system model is a research topic. Over the lifetime, 8697 publications have been published within this topic receiving 259440 citations.


Papers
More filters
Proceedings ArticleDOI
01 Jan 2003
TL;DR: A novel algorithm is presented for embedding a binary image as a watermark into the dc components of the DCT coefficients of 8/spl times/8 blocks that incorporates the feature of texture masking and luminance masking of the human visual system into watermarking.
Abstract: In this paper, we present a novel algorithm for embedding a watermark into a still host image in the DCT domain. Unlike the traditional techniques, we embed a binary image as a watermark into the dc components of the DCT coefficients of 8/spl times/8 blocks. In the proposed algorithm, we incorporate the feature of texture masking and luminance masking of the human visual system into watermarking. The algorithm recovers the watermark without any reference to the original image. Experimental results demonstrate that the visible watermarks embedded with the proposed algorithm are robust to various attacks.

69 citations

Journal ArticleDOI
TL;DR: It is shown that the photoreceptor disarray does not determine the limit to performance for this task, the limit is post-receptoral and can be modelled in terms of a positional uncertainty within the early filters located before the response envelope has been extracted.

69 citations

Journal ArticleDOI
TL;DR: A novel deep generative multiview model for the accurate visual image reconstruction from the human brain activities measured by functional magnetic resonance imaging (fMRI) by using two view-specific generators with a shared latent space is proposed.
Abstract: Neural decoding, which aims to predict external visual stimuli information from evoked brain activities, plays an important role in understanding human visual system. Many existing methods are based on linear models, and most of them only focus on either the brain activity pattern classification or visual stimuli identification. Accurate reconstruction of the perceived images from the measured human brain activities still remains challenging. In this paper, we propose a novel deep generative multiview model for the accurate visual image reconstruction from the human brain activities measured by functional magnetic resonance imaging (fMRI). Specifically, we model the statistical relationships between the two views (i.e., the visual stimuli and the evoked fMRI) by using two view-specific generators with a shared latent space. On the one hand, we adopt a deep neural network architecture for visual image generation, which mimics the stages of human visual processing. On the other hand, we design a sparse Bayesian linear model for fMRI activity generation, which can effectively capture voxel correlations, suppress data noise, and avoid overfitting. Furthermore, we devise an efficient mean-field variational inference method to train the proposed model. The proposed method can accurately reconstruct visual images via Bayesian inference. In particular, we exploit a posterior regularization technique in the Bayesian inference to regularize the model posterior. The quantitative and qualitative evaluations conducted on multiple fMRI data sets demonstrate the proposed method can reconstruct visual images more accurately than the state of the art.

69 citations

Journal ArticleDOI
TL;DR: A new approach to the integration and control of continuously operating visual processes which permits a system to integrate a large number of perceptual processes and to dynamically compose sequences of such processes to perform visual tasks, and illustrated with a system which detects and fixates on different classes of moving objects.

68 citations

Proceedings ArticleDOI
01 Oct 2016
TL;DR: This paper proposes to use human fixation data to train a top-down saliency model that predicts relevant image locations when searching for specific objects and shows that the learned model can successfully prune bounding box proposals without rejecting the ground truth object locations.
Abstract: One of the central tasks for a household robot is searching for specific objects. It does not only require localizing the target object but also identifying promising search locations in the scene if the target is not immediately visible. As computation time and hardware resources are usually limited in robotics, it is desirable to avoid expensive visual processing steps that are exhaustively applied over the entire image. The human visual system can quickly select those image locations that have to be processed in detail for a given task. This allows us to cope with huge amounts of information and to efficiently deploy the limited capacities of our visual system. In this paper, we therefore propose to use human fixation data to train a top-down saliency model that predicts relevant image locations when searching for specific objects. We show that the learned model can successfully prune bounding box proposals without rejecting the ground truth object locations. In this aspect, the proposed model outperforms a model that is trained only on the ground truth segmentations of the target object instead of fixation data.

68 citations


Network Information
Related Topics (5)
Feature (computer vision)
128.2K papers, 1.7M citations
89% related
Feature extraction
111.8K papers, 2.1M citations
86% related
Image segmentation
79.6K papers, 1.8M citations
86% related
Image processing
229.9K papers, 3.5M citations
85% related
Convolutional neural network
74.7K papers, 2M citations
84% related
Performance
Metrics
No. of papers in the topic in previous years
YearPapers
202349
202294
2021279
2020311
2019351
2018348