scispace - formally typeset
Search or ask a question

Showing papers on "Neural coding published in 2012"


Book ChapterDOI
07 Oct 2012
TL;DR: This work extracts semi-local features called random occupancy pattern ROP features, which employ a novel sampling scheme that effectively explores an extremely large sampling space and utilizes a sparse coding approach to robustly encode these features.
Abstract: We study the problem of action recognition from depth sequences captured by depth cameras, where noise and occlusion are common problems because they are captured with a single commodity camera. In order to deal with these issues, we extract semi-local features called random occupancy pattern ROP features, which employ a novel sampling scheme that effectively explores an extremely large sampling space. We also utilize a sparse coding approach to robustly encode these features. The proposed approach does not require careful parameter tuning. Its training is very fast due to the use of the high-dimensional integral image, and it is robust to the occlusions. Our technique is evaluated on two datasets captured by commodity depth cameras: an action dataset and a hand gesture dataset. Our classification results are superior to those obtained by the state of the art approaches on both datasets.

505 citations


Proceedings ArticleDOI
16 Jun 2012
TL;DR: This paper proposes a novel top-down saliency model that jointly learns a Conditional Random Field (CRF) and a discriminative dictionary and proposes a max-margin approach to train the dictionary modulated by CRF, and meanwhile a CRF with sparse coding.
Abstract: Top-down visual saliency facilities object localization by providing a discriminative representation of target objects and a probability map for reducing the search space. In this paper, we propose a novel top-down saliency model that jointly learns a Conditional Random Field (CRF) and a discriminative dictionary. The proposed model is formulated based on a CRF with latent variables. By using sparse codes as latent variables, we train the dictionary modulated by CRF, and meanwhile a CRF with sparse coding. We propose a max-margin approach to train our model via fast inference algorithms. We evaluate our model on the Graz-02 and PASCAL VOC 2007 datasets. Experimental results show that our model performs favorably against the state-of-the-art top-down saliency methods. We also observe that the dictionary update significantly improves the model performance.

350 citations


Journal ArticleDOI
TL;DR: A neuronal coding mechanism in hippocampus is described that can be used to represent the recency of an experience over intervals of hours to days, consistent with behavioral studies showing that the CA1 area is selectively required for temporal coding over such periods.
Abstract: The time when an event occurs can become part of autobiographical memories. In brain structures that support such memories, a neural code should exist that represents when or how long ago events occurred. Here we describe a neuronal coding mechanism in hippocampus that can be used to represent the recency of an experience over intervals of hours to days. When the same event is repeated after such time periods, the activity patterns of hippocampal CA1 cell populations progressively differ with increasing temporal distances. Coding for space and context is nonetheless preserved. Compared with CA1, the firing patterns of hippocampal CA3 cell populations are highly reproducible, irrespective of the time interval, and thus provide a stable memory code over time. Therefore, the neuronal activity patterns in CA1 but not CA3 include a code that can be used to distinguish between time intervals on an extended scale, consistent with behavioral studies showing that the CA1 area is selectively required for temporal coding over such periods.

317 citations


Journal ArticleDOI
TL;DR: A novel dictionary learning method, called Dictionary Learning with Group Sparsity and Graph Regularization (DL-GSGR), where the geometrical structure of atoms is modeled as the graph regularization and the group coherence of learned dictionary can be enforced small enough such that any signal can be group sparse coded effectively.
Abstract: Recently, sparse representation has attracted a lot of interest in various areas. However, the standard sparse representation does not consider the intrinsic structure, i.e., the nonzero elements occur in clusters, called group sparsity. Furthermore, there is no dictionary learning method for group sparse representation considering the geometrical structure of space spanned by atoms. In this paper, we propose a novel dictionary learning method, called Dictionary Learning with Group Sparsity and Graph Regularization (DL-GSGR). First, the geometrical structure of atoms is modeled as the graph regularization. Then, combining group sparsity and graph regularization, the DL-GSGR is presented, which is solved by alternating the group sparse coding and dictionary updating. In this way, the group coherence of learned dictionary can be enforced small enough such that any signal can be group sparse coded effectively. Finally, group sparse representation with DL-GSGR is applied to 3-D medical image denoising and image fusion. Specifically, in 3-D medical image denoising, a 3-D processing mechanism (using the similarity among nearby slices) and temporal regularization (to perverse the correlations across nearby slices) are exploited. The experimental results on 3-D image denoising and image fusion demonstrate the superiority of our proposed denoising and fusion approaches.

310 citations


Proceedings Article
Ren Xiaofeng1, Liefeng Bo1
03 Dec 2012
TL;DR: This work shows that contour detection accuracy can be significantly improved by computing Sparse Code Gradients (SCG), which measure contrast using patch representations automatically learned through sparse coding, which is verified on the NYU Depth Dataset.
Abstract: Finding contours in natural images is a fundamental problem that serves as the basis of many tasks such as image segmentation and object recognition. At the core of contour detection technologies are a set of hand-designed gradient features, used by most approaches including the state-of-the-art Global Pb (gPb) operator. In this work, we show that contour detection accuracy can be significantly improved by computing Sparse Code Gradients (SCG), which measure contrast using patch representations automatically learned through sparse coding. We use K-SVD for dictionary learning and Orthogonal Matching Pursuit for computing sparse codes on oriented local neighborhoods, and apply multi-scale pooling and power transforms before classifying them with linear SVMs. By extracting rich representations from pixels and avoiding collapsing them prematurely, Sparse Code Gradients effectively learn how to measure local contrasts and find contours. We improve the F-measure metric on the BSDS500 benchmark to 0.74 (up from 0.71 of gPb contours). Moreover, our learning approach can easily adapt to novel sensor data such as Kinect-style RGB-D cameras: Sparse Code Gradients on depth maps and surface normals lead to promising contour detection using depth and depth+color, as verified on the NYU Depth Dataset.

269 citations


Journal ArticleDOI
TL;DR: Important findings regarding oscillations in primary motor cortex, synchronization between cortex and spinal cord, synchronization Between cortical regions, as well as abnormal synchronization patterns in a selection of motor dysfunctions are highlighted.
Abstract: Synchronization of neural activity is considered essential for information processing in the nervous system. Both local and inter-regional synchronization are omnipresent in different frequency regimes and relate to a variety of behavioral and cognitive functions. Over the years, many studies have sought to elucidate the question how alpha/mu, beta, and gamma synchronization contribute to motor control. Here, we review these studies with the purpose to delineate what they have added to our understanding of the neural control of movement. We highlight important findings regarding oscillations in primary motor cortex, synchronization between cortex and spinal cord, synchronization between cortical regions, as well as abnormal synchronization patterns in a selection of motor dysfunctions. The interpretation of synchronization patterns benefits from combining results of invasive and non-invasive recordings, different data analysis tools, and modeling work. Importantly, although synchronization is deemed to play a vital role, it is not the only mechanism for neural communication. Spike timing and rate coding act together during motor control and should therefore both be accounted for when interpreting movement-related activity.

229 citations


Journal ArticleDOI
TL;DR: The sparse coding method for satellite scene classification is introduced, a two-stage linear support vector machine (SVM) classifier is designed and an improved rotation invariant texture descriptor based on LTPs is presented.
Abstract: This article presents a new method for high-resolution satellite scene classification. Specifically, we make three main contributions: (1) we introduce the sparse coding method for satellite scene classification; (2) we present local ternary pattern histogram Fourier (LTP-HF) features, an improved rotation invariant texture descriptor based on LTPs; (3) we effectively combine a set of diverse and complementary features to further improve the performance. A two-stage linear support vector machine (SVM) classifier is designed for this purpose. In the first stage, the SVM is used to generate probability images with a scale invariant feature transform (SIFT), LTP-HF and colour histogram features, respectively. The generated probability images with different features are fused in the second stage in order to obtain the final classification results. Experimental results show that the suggested classification method achieves very promising performance.

223 citations


Journal ArticleDOI
21 Jun 2012-Neuron
TL;DR: It is shown that odors evoke transient bursts locked to sniff onset and that odor identity can be better decoded using burst spike counts than by spike latencies or temporal patterns and that behavioral performance limits arise downstream of aPC.

215 citations


Journal ArticleDOI
Hao Sun, Xian Sun, Hongqi Wang, Yu Li, Xiangjuan Li 
TL;DR: This letter proposes a new detection framework based on spatial sparse coding bag-of-words (BOW) (SSCBOW) model, which not only represents the relative position of the parts of a target but also has the ability to handle rotation variations.
Abstract: Automatic detection for targets with complex shape in high-resolution remote sensing images is a challenging task. In this letter, we propose a new detection framework based on spatial sparse coding bag-of-words (BOW) (SSCBOW) model to solve this problem. Specifically, after selecting a processing unit by the sliding window and extracting features, a new spatial mapping strategy is used to encode the geometric information, which not only represents the relative position of the parts of a target but also has the ability to handle rotation variations. Moreover, instead of K-means for visual-word encoding in the traditional BOW model, sparse coding is introduced to achieve a much lower reconstruction error. Finally, the SSCBOW representation is combined with linear support vector machine for target detection. The experimental results demonstrate the precision and robustness of our detection method based on the SSCBOW model.

194 citations


Journal ArticleDOI
TL;DR: This work shows that the combination of a large, dense multielectrode array and a novel, mostly automated spike-sorting algorithm allowed them to record simultaneously from a highly overlapping population of >200 ganglion cells in the salamander retina, allowing unprecedented access to the complete neural representation of visual information.
Abstract: Recording simultaneously from essentially all of the relevant neurons in a local circuit is crucial to understand how they collectively represent information. Here we show that the combination of a large, dense multielectrode array and a novel, mostly automated spike-sorting algorithm allowed us to record simultaneously from a highly overlapping population of >200 ganglion cells in the salamander retina. By combining these methods with labeling and imaging, we showed that up to 95% of the ganglion cells over the area of the array were recorded. By measuring the coverage of visual space by the receptive fields of the recorded cells, we concluded that our technique captured a neural population that forms an essentially complete representation of a region of visual space. This completeness allowed us to determine the spatial layout of different cell types as well as identify a novel group of ganglion cells that responded reliably to a set of naturalistic and artificial stimuli but had no measurable receptive field. Thus, our method allows unprecedented access to the complete neural representation of visual information, a crucial step for the understanding of population coding in sensory systems.

184 citations


Journal ArticleDOI
TL;DR: This work considers networks of general integrate-and-fire cells with arbitrary architecture, and derives explicit expressions for the approximate cross-correlation between constituent cells, which help to identify the important factors that shape coordinated neural activity in such networks.
Abstract: Novel experimental techniques reveal the simultaneous activity of larger and larger numbers of neurons. As a result there is increasing interest in the structure of cooperative – or correlated – activity in neural populations, and in the possible impact of such correlations on the neural code. A fundamental theoretical challenge is to understand how the architecture of network connectivity along with the dynamical properties of single cells shape the magnitude and timescale of correlations. We provide a general approach to this problem by extending prior techniques based on linear response theory. We consider networks of general integrate-and-fire cells with arbitrary architecture, and provide explicit expressions for the approximate cross-correlation between constituent cells. These correlations depend strongly on the operating point (input mean and variance) of the neurons, even when connectivity is fixed. Moreover, the approximations admit an expansion in powers of the matrices that describe the network architecture. This expansion can be readily interpreted in terms of paths between different cells. We apply our results to large excitatory-inhibitory networks, and demonstrate first how precise balance – or lack thereof – between the strengths and timescales of excitatory and inhibitory synapses is reflected in the overall correlation structure of the network. We then derive explicit expressions for the average correlation structure in randomly connected networks. These expressions help to identify the important factors that shape coordinated neural activity in such networks.

Book ChapterDOI
07 Oct 2012
TL;DR: In this paper, the Stein kernel is used to embed Riemannian manifolds into reproducing kernel Hilbert spaces, which leads to a convex and kernel version of the Lasso problem, which can be solved efficiently.
Abstract: Recent advances suggest that a wide range of computer vision problems can be addressed more appropriately by considering non-Euclidean geometry. This paper tackles the problem of sparse coding and dictionary learning in the space of symmetric positive definite matrices, which form a Riemannian manifold. With the aid of the recently introduced Stein kernel (related to a symmetric version of Bregman matrix divergence), we propose to perform sparse coding by embedding Riemannian manifolds into reproducing kernel Hilbert spaces. This leads to a convex and kernel version of the Lasso problem, which can be solved efficiently. We furthermore propose an algorithm for learning a Riemannian dictionary (used for sparse coding), closely tied to the Stein kernel. Experiments on several classification tasks (face recognition, texture classification, person re-identification) show that the proposed sparse coding approach achieves notable improvements in discrimination accuracy, in comparison to state-of-the-art methods such as tensor sparse coding, Riemannian locality preserving projection, and symmetry-driven accumulation of local features.

Proceedings ArticleDOI
16 Jun 2012
TL;DR: A simple yet effective algorithm based upon the sparse representation of natural scene statistics (NSS) feature that outperforms representative BIQA algorithms and some full-reference metrics is introduced.
Abstract: Blind image quality assessment (BIQA) is an important yet difficult task in image processing related applications. Existing algorithms for universal BIQA learn a mapping from features of an image to the corresponding subjective quality or divide the image into different distortions before mapping. Although these algorithms are promising, they face the following problems: 1) they require a large number of samples (pairs of distorted image and its subjective quality) to train a robust mapping; 2) they are sensitive to different datasets; and 3) they have to be retrained when new training samples are available. In this paper, we introduce a simple yet effective algorithm based upon the sparse representation of natural scene statistics (NSS) feature. It consists of three key steps: extracting NSS features in the wavelet domain, representing features via sparse coding, and weighting differential mean opinion scores by the sparse coding coefficients to obtain the final visual quality values. Thorough experiments on standard databases show that the proposed algorithm outperforms representative BIQA algorithms and some full-reference metrics.

Journal ArticleDOI
23 Aug 2012-Neuron
TL;DR: Evidence is reviewed for a bias in neuronal circuits toward temporal coding and the coexistence of rate and temporal coding during population rhythm generation and the coincident expression of multiple types of gamma rhythm in sensory cortex suggests a mechanistic substrate for combining rate andporal codes on the basis of stimulus strength.

Journal ArticleDOI
TL;DR: This study presents a methodological approach employing magnetoencephalography (MEG) and machine learning techniques to investigate the flow of perceptual and semantic information decodable from neural activity in the half second during which the brain comprehends the meaning of a concrete noun.

Journal ArticleDOI
21 Jun 2012-Neuron
TL;DR: Adaptive shape sampling is used to demonstrate explicit coding of medial axis shape in high-level object cortex (macaque monkey inferotemporal cortex or IT), and metric shape analyses revealed a coding continuum, along which most neurons represent a configuration of both medial axis and surface components.

Journal ArticleDOI
TL;DR: This work focuses on emerging evidence that value coding in a number of brain areas is context dependent, varying as a function of both the current choice set and previously experienced values.
Abstract: To survive in a dynamic environment, an organism must be able to effectively learn, store, and recall the expected benefits and costs of potential actions. The nature of the valuation and decision processes is thus of fundamental interest to researchers at the intersection of psychology, neuroscience, and economics. Although normative theories of choice have outlined the theoretical structure of these valuations, recent experiments have begun to reveal how value is instantiated in the activity of neurons and neural circuits. Here, we review the various forms of value coding that have been observed in different brain systems and examine the implications of these value representations for both neural circuits and behavior. In particular, we focus on emerging evidence that value coding in a number of brain areas is context dependent, varying as a function of both the current choice set and previously experienced values. Similar contextual modulation occurs widely in the sensory system, and efficient coding principles derived in the sensory domain suggest a new framework for understanding the neural coding of value.

Journal ArticleDOI
TL;DR: This work focuses on the approach to resampling known as jitter, and relies on an intuitive and rigorous statistical framework known as conditional modeling to reveal otherwise hidden assumptions and to support precise conclusions.
Abstract: The existence and role of fine-temporal structure in the spiking activity of central neurons is the subject of an enduring debate among physiologists. To a large extent, the problem is a statistical one: what inferences can be drawn from neurons monitored in the absence of full control over their presynaptic environments? In principle, properly crafted resampling methods can still produce statistically correct hypothesis tests. We focus on the approach to resampling known as jitter. We review a wide range of jitter techniques, illustrated by both simulation experiments and selected analyses of spike data from motor cortical neurons. We rely on an intuitive and rigorous statistical framework known as conditional modeling to reveal otherwise hidden assumptions and to support precise conclusions. Among other applications, we review statistical tests for exploring any proposed limit on the rate of change of spiking probabilities, exact tests for the significance of repeated fine-temporal patterns of spikes, and the construction of acceptance bands for testing any purported relationship between sensory or motor variables and synchrony or other fine-temporal events.

Journal ArticleDOI
TL;DR: A novel and unique prediction of efficient coding, the relationships between projection patterns of individual cones to all ganglion cells, was consistent with the observed projection patterns in the retina, indicating a high level of efficiency with near-optimal redundancy in visual signaling by the retina.
Abstract: Sensory neurons have been hypothesized to efficiently encode signals from the natural environment subject to resource constraints. The predictions of this efficient coding hypothesis regarding the spatial filtering properties of the visual system have been found consistent with human perception, but they have not been compared directly with neural responses. Here, we analyze the information that retinal ganglion cells transmit to the brain about the spatial information in natural images subject to three resource constraints: the number of retinal ganglion cells, their total response variances, and their total synaptic strengths. We derive a model that optimizes the transmitted information and compare it directly with measurements of complete functional connectivity between cone photoreceptors and the four major types of ganglion cells in the primate retina, obtained at single-cell resolution. We find that the ganglion cell population exhibited 80% efficiency in transmitting spatial information relative to the model. Both the retina and the model exhibited high redundancy (∼30%) among ganglion cells of the same cell type. A novel and unique prediction of efficient coding, the relationships between projection patterns of individual cones to all ganglion cells, was consistent with the observed projection patterns in the retina. These results indicate a high level of efficiency with near-optimal redundancy in visual signaling by the retina.

Journal ArticleDOI
TL;DR: The evidence for and against the idea that a synchrony code defines, at least in part, the input–output function between the cerebellar cortex and nuclei are explored.
Abstract: The cerebellum regulates complex movements and is also implicated in cognitive tasks, and cerebellar dysfunction is consequently associated not only with movement disorders, but also with conditions like autism and dyslexia. How information is encoded by specific cerebellar firing patterns remains debated, however. A central question is how the cerebellar cortex transmits its integrated output to the cerebellar nuclei via GABAergic synapses from Purkinje neurons. Possible answers come from accumulating evidence that subsets of Purkinje cells synchronize their firing during behaviors that require the cerebellum. Consistent with models predicting that coherent activity of inhibitory networks has the capacity to dictate firing patterns of target neurons, recent experimental work supports the idea that inhibitory synchrony may regulate the response of cerebellar nuclear cells to Purkinje inputs, owing to the interplay between unusually fast inhibitory synaptic responses and high rates of intrinsic activity. Data from multiple laboratories lead to a working hypothesis that synchronous inhibitory input from Purkinje cells can set the timing and rate of action potentials produced by cerebellar nuclear cells, thereby relaying information out of the cerebellum. If so, then changing spatiotemporal patterns of Purkinje activity would allow different subsets of inhibitory neurons to control cerebellar output at different times. Here we explore the evidence for and against the idea that a synchrony code defines, at least in part, the input–output function between the cerebellar cortex and nuclei. We consider the literature on the existence of simple spike synchrony, convergence of Purkinje neurons onto nuclear neurons, and intrinsic properties of nuclear neurons that contribute to responses to inhibition. Finally, we discuss factors that may disrupt or modulate a synchrony code and describe the potential contributions of inhibitory synchrony to other motor circuits.

Journal ArticleDOI
TL;DR: A set of piecewise linear spiking neuron models, which can reproduce different behaviors, similar to the biological neuron, both for a single neuron as well as a network of neurons are presented.
Abstract: There has been a strong push recently to examine biological scale simulations of neuromorphic algorithms to achieve stronger inference capabilities. This paper presents a set of piecewise linear spiking neuron models, which can reproduce different behaviors, similar to the biological neuron, both for a single neuron as well as a network of neurons. The proposed models are investigated, in terms of digital implementation feasibility and costs, targeting large scale hardware implementation. Hardware synthesis and physical implementations on FPGA show that the proposed models can produce precise neural behaviors with higher performance and considerably lower implementation costs compared with the original model. Accordingly, a compact structure of the models which can be trained with supervised and unsupervised learning algorithms has been developed. Using this structure and based on a spike rate coding, a character recognition case study has been implemented and tested.

Journal ArticleDOI
TL;DR: Compositional coding in rule representation shows that the code used to store rule information in prefrontal cortex is compositional, and suggests that it might be possible to decode other complex action plans by learning the neural patterns of the known composing elements.
Abstract: Rules are widely used in everyday life to organize actions and thoughts in accordance with our internal goals. At the simplest level, single rules can be used to link individual sensory stimuli to their appropriate responses. However, most tasks are more complex and require the concurrent application of multiple rules. Experiments on humans and monkeys have shown the involvement of a frontoparietal network in rule representation. Yet, a fundamental issue still needs to be clarified: Is the neural representation of multiple rules compositional, that is, built on the neural representation of their simple constituent rules? Subjects were asked to remember and apply either simple or compound rules. Multivariate decoding analyses were applied to functional magnetic resonance imaging data. Both ventrolateral frontal and lateral parietal cortex were involved in compound representation. Most importantly, we were able to decode the compound rules by training classifiers only on the simple rules they were composed of. This shows that the code used to store rule information in prefrontal cortex is compositional. Compositional coding in rule representation suggests that it might be possible to decode other complex action plans by learning the neural patterns of the known composing elements.

Journal ArticleDOI
TL;DR: Heterogeneity decreases the threshold for synchronization, and its strength is nonlinearly related to the network mean firing rate, and conditions are shown under which heterogeneity optimizes network information transmission for either temporal or rate coding.
Abstract: The effect of cellular heterogeneity on the coding properties of neural populations is studied analytically and numerically. We find that heterogeneity decreases the threshold for synchronization, and its strength is nonlinearly related to the network mean firing rate. In addition, conditions are shown under which heterogeneity optimizes network information transmission for either temporal or rate coding, with high input frequencies leading to different effects for each coding strategy. The results are shown to be robust for more realistic conditions.

Journal ArticleDOI
TL;DR: In both the retina and the LGN, relative spike times are more precise, less affected by pre-landing history and global contrast than absolute ones, and lead to robust contrast invariant orientation representations in V1.
Abstract: We have built a phenomenological spiking model of the cat early visual system comprising the retina, the Lateral Geniculate Nucleus (LGN) and V1's layer 4, and established four main results (1) When exposed to videos that reproduce with high fidelity what a cat experiences under natural conditions, adjacent Retinal Ganglion Cells (RGCs) have spike-time correlations at a short timescale (~30 ms), despite neuronal noise and possible jitter accumulation. (2) In accordance with recent experimental findings, the LGN filters out some noise. It thus increases the spike reliability and temporal precision, the sparsity, and, importantly, further decreases down to ~15 ms adjacent cells' correlation timescale. (3) Downstream simple cells in V1's layer 4, if equipped with Spike Timing-Dependent Plasticity (STDP), may detect these fine-scale cross-correlations, and thus connect principally to ON- and OFF-centre cells with Receptive Fields (RF) aligned in the visual space, and thereby become orientation selective, in accordance with Hubel and Wiesel (Journal of Physiology 160:106---154, 1962) classic model. Up to this point we dealt with continuous vision, and there was no absolute time reference such as a stimulus onset, yet information was encoded and decoded in the relative spike times. (4) We then simulated saccades to a static image and benchmarked relative spike time coding and time-to-first spike coding w.r.t. to saccade landing in the context of orientation representation. In both the retina and the LGN, relative spike times are more precise, less affected by pre-landing history and global contrast than absolute ones, and lead to robust contrast invariant orientation representations in V1.

Journal ArticleDOI
TL;DR: It is concluded that the analysis of the output of spinal motor neurons from muscle signals provides a unique means for understanding the neural coding of movement in vivo in humans and thus for reproducing this code artificially with the aim of restoring lost or impaired motor functions.
Abstract: This review describes methods for interfacing motor neurons from muscle recordings and their applications in studies on the neural control of movement and in the design of technologies for neurorehabilitation. After describing methods for accessing the neural drive to muscles in vivo in humans, we discuss the mechanisms of transmission of synaptic input into motor neuron output and of force generation. The synaptic input received by a motor neuron population is largely common among motor neurons. This allows linear transmission of the input and a reduced dimensionality of control by the central nervous system. Force is generated by low-pass filtering the neural signal sent to the muscle. These concepts on neural control of movement are used for the development of neurorehabilitation technologies, which are discussed with representative examples on movement replacement, restoration, and neuromodulation. It is concluded that the analysis of the output of spinal motor neurons from muscle signals provides a unique means for understanding the neural coding of movement in vivo in humans and thus for reproducing this code artificially with the aim of restoring lost or impaired motor functions.

Proceedings ArticleDOI
Xiaoqiang Lu1, Haoliang Yuan1, Pingkun Yan1, Yuan Yuan1, Xuelong Li1 
16 Jun 2012
TL;DR: A novel sparse coding method is proposed to preserve the geometrical structure of the dictionary and the sparse coefficients of the data and can preserve the incoherence of dictionary entries, which is critical for sparse representation.
Abstract: The choice of the over-complete dictionary that sparsely represents data is of prime importance for sparse coding-based image super-resolution. Sparse coding is a typical unsupervised learning method to generate an over-complete dictionary. However, most of the sparse coding methods for image super-resolution fail to simultaneously consider the geometrical structure of the dictionary and corresponding coefficients, which may result in noticeable super-resolution reconstruction artifacts. In this paper, a novel sparse coding method is proposed to preserve the geometrical structure of the dictionary and the sparse coefficients of the data. Moreover, the proposed method can preserve the incoherence of dictionary entries, which is critical for sparse representation. Inspired by the development on non-local self-similarity and manifold learning, the proposed sparse coding method can provide the sparse coefficients and learned dictionary from a new perspective, which have both reconstruction and discrimination properties to enhance the learning performance. Extensive experimental results on image super-resolution have demonstrated the effectiveness of the proposed method.

Book ChapterDOI
07 Oct 2012
TL;DR: An intermediate representation for deformable part models is developed and it is shown that this representation has favorable performance characteristics for multi-class problems when the number of classes is high and is well suited to a parallel implementation.
Abstract: We develop an intermediate representation for deformable part models and show that this representation has favorable performance characteristics for multi-class problems when the number of classes is high. Our model uses sparse coding of part filters to represent each filter as a sparse linear combination of shared dictionary elements. This leads to a universal set of parts that are shared among all object classes. Reconstruction of the original part filter responses via sparse matrix-vector product reduces computation relative to conventional part filter convolutions. Our model is well suited to a parallel implementation, and we report a new GPU DPM implementation that takes advantage of sparse coding of part filters. The speed-up offered by our intermediate representation and parallel computation enable real-time DPM detection of 20 different object classes on a laptop computer.

Journal ArticleDOI
12 Jul 2012-Neuron
TL;DR: Using 2-photon in vivo calcium imaging to measure responses of networks of neurons in primary somatosensory cortex, it is discovered that associative fear learning, in which whisker stimulation is paired with foot shock, enhances sparse population coding and robustness of the conditional stimulus, yet decreases total network activity.

Proceedings ArticleDOI
16 Jun 2012
TL;DR: This work presents a new general sparse coding model that relates signals from the two spaces by their sparse representations and the corresponding dictionaries, and tailor the general model to learning dictionaries for compressive sensing recovery and single image super-resolution to demonstrate its effectiveness.
Abstract: In this paper, we propose a bilevel sparse coding model for coupled feature spaces, where we aim to learn dictionaries for sparse modeling in both spaces while enforcing some desired relationships between the two signal spaces. We first present our new general sparse coding model that relates signals from the two spaces by their sparse representations and the corresponding dictionaries. The learning algorithm is formulated as a generic bilevel optimization problem, which is solved by a projected first-order stochastic gradient descent algorithm. This general sparse coding model can be applied to many specific applications involving coupled feature spaces in computer vision and signal processing. In this work, we tailor our general model to learning dictionaries for compressive sensing recovery and single image super-resolution to demonstrate its effectiveness. In both cases, the new sparse coding model remarkably outperforms previous approaches in terms of recovery accuracy.

Journal ArticleDOI
TL;DR: The temporal coding improvement—responses more precisely temporally following a stimulus when animals were required to attend to it—expands the framework of possible mechanisms of attention to include increasing temporal precision of stimulus following.
Abstract: The effect of attention on single neuron responses in the auditory system is unresolved We found that when monkeys discriminated temporally amplitude modulated (AM) from unmodulated sounds, primary auditory cortical (A1) neurons better discriminated those sounds than when the monkeys were not discriminating them This was observed for both average firing rate and vector strength (VS), a measure of how well neurons temporally follow the stimulus' temporal modulation When data were separated by nonsynchronized and synchronized responses, the firing rate of nonsynchronized responses best distinguished AM- noise from unmodulated noise, followed by VS for synchronized responses, with firing rate for synchronized neurons providing the poorest AM discrimination Firing rate-based AM discrimination for synchronized neurons, however, improved most with task engagement, showing that the least sensitive code in the passive condition improves the most with task engagement Rate coding improved due to larger increases in absolute firing rate at higher modulation depths than for lower depths and unmodulated sounds Relative to spontaneous activity (which increased with engagement), the response to unmodulated sounds decreased substantially The temporal coding improvement—responses more precisely temporally following a stimulus when animals were required to attend to it—expands the framework of possible mechanisms of attention to include increasing temporal precision of stimulus following These findings provide a crucial step to understanding the coding of temporal modulation and support a model in which rate and temporal coding work in parallel, permitting a multiplexed code for temporal modulation, and for a complementary representation of rate and temporal coding