scispace - formally typeset
Proceedings ArticleDOI

Phrase bigrams for continuous speech recognition

E.P. Giachin
- Vol. 1, pp 225-228
Reads0
Chats0
TLDR
Two procedures for automatically determining frequent phrases (within the framework of a probabilistic language model) in an unlabeled training set of written sentences are discussed and one procedure is optimal since it minimises the set perplexity.
Abstract
In some speech recognition tasks, such as man-machine dialogue systems, the spoken sentences include several recurrent phrases. A bigram language model does not adequately represent these phrases because it underestimates their probability. A better approach consists of modeling phrases as if they were individual dictionary elements. They we inserted as additional entries into the word lexicon, on which bigrams are finally computed. This paper discusses two procedures for automatically determining frequent phrases (within the framework of a probabilistic language model) in an unlabeled training set of written sentences. One procedure is optimal since it minimises the set perplexity. The other, based on information theoretic criteria, insures that the resulting model has a high statistical robustness. The two procedures are tested on a 762-word spontaneous speech recognition task. They give similar results and provide a moderate improvement over standard bigrams.

read more

Citations
More filters
Journal ArticleDOI

How may I help you

TL;DR: This paper focuses on the task of automatically routing telephone calls based on a user's fluently spoken response to the open-ended prompt of “ How may I help you? ”.
Journal ArticleDOI

Grammar fragment acquisition using syntactic and semantic clustering

TL;DR: A method and apparatus are provided for automatically acquiring grammar fragments for recognizing and understanding fluently spoken language.
Journal ArticleDOI

Toward a unified approach to statistical language modeling for Chinese

TL;DR: This article presents a unified approach to Chinese statistical language modeling, which automatically and consistently gathers a high-quality training data set from the Web, creates ahigh-quality lexicon, segments the training data using this Lexicon, and compresses the language model by using the maximum likelihood principle, which is consistent with trigram model training.
Posted Content

Metrics for evaluating dialogue strategies in a spoken language system

TL;DR: A new metric is defined, the implicit recovery, which allows to measure the ability of a dialogue manager to deal with errors by different levels of analysis and is compared to two different approaches to dialogue repair strategies.
Proceedings ArticleDOI

Improving speech understanding by incorporating database constraints and dialogue history

TL;DR: This work shows how knowledge sources in the course of a (man-machine) dialogue may be utilized in a stochastic framework to improve speech understanding.
References
More filters
Proceedings ArticleDOI

On smoothing techniques for bigram-based natural language modelling

TL;DR: It is shown that the leaving-one-out method in combination with the maximum likelihood criterion can be efficiently used for the optimal estimation of interpolation parameters.
Proceedings Article

Towards better language models for spontaneous speech.

Bernhard Suhm, +1 more
TL;DR: Several methods to improve the language models of the speech decoder of thespeech translation system for spontaneous spoken dialogs attempt to take advantage of natural equivalence word classes, frequently occur-ing word phrases, and discourse structure.