scispace - formally typeset
Journal ArticleDOI

Deep-Sparse-Representation-Based Features for Speech Recognition

Reads0
Chats0
TLDR
This paper proposes to use a multilevel decomposition (having multiple layers), also known as the deep sparse representation (DSR), to derive a feature representation for speech recognition, and reveals that the representations obtained at different sparse layers of the proposed DSR model have complimentary information.
Abstract
Features derived using sparse representation (SR)-based approaches have been shown to yield promising results for speech recognition tasks. In most of the approaches, the SR corresponding to speech signal is estimated using a dictionary, which could be either exemplar based or learned. However, a single-level decomposition may not be suitable for the speech signal, as it contains complex hierarchical information about various hidden attributes. In this paper, we propose to use a multilevel decomposition (having multiple layers), also known as the deep sparse representation (DSR), to derive a feature representation for speech recognition. Instead of having a series of sparse layers, the proposed framework employs a dense layer between two sparse layers, which helps in efficient implementation. Our studies reveal that the representations obtained at different sparse layers of the proposed DSR model have complimentary information. Thus, the final feature representation is derived after concatenating the representations obtained at the sparse layers. This results in a more discriminative representation, and improves the speech recognition performance. Since the concatenation results in a high-dimensional feature, principal component analysis is used to reduce the dimension of the obtained feature. Experimental studies demonstrate that the proposed feature outperforms existing features for various speech recognition tasks.

read more

Citations
More filters
Journal ArticleDOI

A sparse deep belief network with efficient fuzzy learning framework.

TL;DR: As a novel cross-model, SDBFNN combines the advantages of both pre-training technique and fuzzy neural network to improve modeling capability and achieves better performance than the existing methods in learning speed, modeling accuracy and robustness.
Journal ArticleDOI

Multiview Concept Learning Via Deep Matrix Factorization

TL;DR: This work presents the deep multiview concept learning (DMCL) method, which hierarchically factorizes the multIView data, and tries to explicitly model consistent and complementary information and capture semantic structures at the highest abstraction level.
Journal ArticleDOI

Weighted discriminative collaborative competitive representation for robust image classification

TL;DR: The proposed WDCCR designs the discriminative and competitive collaborative representation among all the classes by fully considering the class information, and introduces the constraint of the weighted categorical representation coefficients into the proposed model for further enhancing the power of discriminatives and competitive representation.
Proceedings ArticleDOI

Deep Multi-View Concept Learning

TL;DR: DMCL performs nonnegative factorization of the data hierarchically, and tries to capture semantic structures and explicitly model consistent and complementary information in multi-view data at the highest abstraction level, and develops a block coordinate descent algorithm for DMCL.
Journal ArticleDOI

Classification of multiclass motor imagery EEG signal using sparsity approach

TL;DR: A sparse representation based classification technique has been proposed to classify multi-tasks MI electroencephalogram data and the results substantiate that the proposed sparsity approach performs significantly better than the existing classifiers.
References
More filters
Journal Article

Visualizing Data using t-SNE

TL;DR: A new technique called t-SNE that visualizes high-dimensional data by giving each datapoint a location in a two or three-dimensional map, a variation of Stochastic Neighbor Embedding that is much easier to optimize, and produces significantly better visualizations by reducing the tendency to crowd points together in the center of the map.

Statistical learning theory

TL;DR: Presenting a method for determining the necessary and sufficient conditions for consistency of learning process, the author covers function estimates from small data pools, applying these estimations to real-life problems, and much more.
Journal ArticleDOI

Robust Face Recognition via Sparse Representation

TL;DR: This work considers the problem of automatically recognizing human faces from frontal views with varying expression and illumination, as well as occlusion and disguise, and proposes a general classification algorithm for (image-based) object recognition based on a sparse representation computed by C1-minimization.
Journal ArticleDOI

Deep Neural Networks for Acoustic Modeling in Speech Recognition: The Shared Views of Four Research Groups

TL;DR: This article provides an overview of progress and represents the shared views of four research groups that have had recent successes in using DNNs for acoustic modeling in speech recognition.
Journal ArticleDOI

$rm K$ -SVD: An Algorithm for Designing Overcomplete Dictionaries for Sparse Representation

TL;DR: A novel algorithm for adapting dictionaries in order to achieve sparse signal representations, the K-SVD algorithm, an iterative method that alternates between sparse coding of the examples based on the current dictionary and a process of updating the dictionary atoms to better fit the data.
Related Papers (5)