scispace - formally typeset
Proceedings ArticleDOI

Neural code-excited linear prediction for low power speech compression

S. Kamarsu, +1 more
- Vol. 2, pp 415-420
TLDR
This paper discusses the use of artificial neural learning methods for low bit-rate speech compression, potentially in non-stationary environments and employs two unsupervised learning algorithms: frequency-sensitive competitive learning and Kohonen's self-organizing maps.
Abstract
In this paper, we discuss the use of artificial neural learning methods for low bit-rate speech compression, potentially in non-stationary environments. Unsupervised learning algorithms are particularly well-suited for vector quantization (VQ) which is used in many speech compression applications. We discuss two unsupervised learning algorithms: frequency-sensitive competitive learning and Kohonen's self-organizing maps which have both been investigated for learning the codebook vectors in an adaptive vector quantizer. In contrast with earlier work, we have employed these learning rules in VQ of the linear predictive coding (LPC) prediction residual. The performance of these unsupervised learning algorithms in speaker-dependent and speaker-independent speech compression are presented. Our results compare favourably with those of code-excited linear prediction (CELP) requiring reduced computational power with a tolerable reduction in speech quality. We also explore the effects of limited precision on classification and learning in competitive learning algorithms for low power VLSI implementations.

read more

Citations
More filters
Proceedings ArticleDOI

Adaptive compression of animated sequences

TL;DR: This investigation indicates that a compression ratio of better than 100:1 is a reasonable expectation for highly repetitious video signals such as those found in animated cartoons.
References
More filters
Journal Article

Vector quantization

TL;DR: During the past few years several design algorithms have been developed for a variety of vector quantizers and the performance of these codes has been studied for speech waveforms, speech linear predictive parameter vectors, images, and several simulated random processes.
Proceedings ArticleDOI

Code-excited linear prediction(CELP): High-quality speech at very low bit rates

TL;DR: A code-excited linear predictive coder in which the optimum innovation sequence is selected from a code book of stored sequences to optimize a given fidelity criterion, indicating that a random code book has a slight speech quality advantage at low bit rates.
Journal ArticleDOI

Competitive learning algorithms for vector quantization

TL;DR: A new competitive-learning algorithm based on the “conscience” learning method is introduced that is shown to be efficient and yields near-optimal results in vector quantization for data compression.
Journal ArticleDOI

Neural networks for vector quantization of speech and images

TL;DR: The authors show how a collection of neural units can be used efficiently for VQ encoding, with the units performing the bulk of the computation in parallel, and describe two unsupervised neural network learning algorithms for training the vector quantizer.
Journal ArticleDOI

Pitch prediction filters in speech coding

TL;DR: It is found that the F-P cascade (formant filter before the pitch filter) outperforms the P-F cascade for the both transversal- and lattice-structured predictors.
Related Papers (5)