scispace - formally typeset
Search or ask a question
Topic

Word error rate

About: Word error rate is a research topic. Over the lifetime, 11939 publications have been published within this topic receiving 298031 citations.


Papers
More filters
Journal ArticleDOI
TL;DR: Experimental results obtained on the IAM off-line database demonstrate that consistent word error rate reductions can be achieved with neural network language models when compared with statistical N-gram language models on the three tested systems.

72 citations

Journal ArticleDOI
TL;DR: The proposed support vector machine classifier for broken bar detection in electrical induction machine is a reliable online method, which has high robustness to load variations and changing operating conditions and is suitable for use in real-time online applications in industrial drives.
Abstract: Highlights? System for broken bar detection for wide slip range. ? Detection based on measuring only a motor current. ? No need for mathematical models, classifier is trained on acquired data. ? Detection of broken bar at low slip, where classical broken bar detection classifiers are not applicable. ? Reliable, mobile, and cost effective system that can be successfully applied in real working environment. This paper presents a support vector machine classifier for broken bar detection in electrical induction machine. It is a reliable online method, which has high robustness to load variations and changing operating conditions. The phase current is only physical value to be measured. The steady state current is analyzed for broken bar fault via motor current signature analysis technique based on Hilbert transform. A two dimensional feature space is proposed. The features are: magnitude and frequency of characteristic peak extracted from spectrum of Hilbert transform series of the phase current. For classification task support vector machine is used due to its good robustness and generalization performances. A comparative analysis of linear, Gaussian and quadratic kernel function versus error rate and number of support vectors is done. The proposed classifier successfully detects a broken bar in various operational situations. The proposed method is sufficiently accurate, fast, and robust to load changes, which makes it suitable for use in real-time online applications in industrial drives.

72 citations

Journal ArticleDOI
TL;DR: Novel unsupervised frequency domain and cepstral domain equalizations that increase ASR resistance to LE are proposed and incorporated in a recognition scheme employing a codebook of noisy acoustic models and provide an absolute word error rate reduction on 10-dB signal-to-noise ratio data.
Abstract: In the presence of environmental noise, speakers tend to adjust their speech production in an effort to preserve intelligible communication. The noise-induced speech adjustments, called Lombard effect (LE), are known to severely impact the accuracy of automatic speech recognition (ASR) systems. The reduced performance results from the mismatch between the ASR acoustic models trained typically on noise-clean neutral (modal) speech and the actual parameters of noisy LE speech. In this study, novel unsupervised frequency domain and cepstral domain equalizations that increase ASR resistance to LE are proposed and incorporated in a recognition scheme employing a codebook of noisy acoustic models. In the frequency domain, short-time speech spectra are transformed towards neutral ASR acoustic models in a maximum-likelihood fashion. Simultaneously, dynamics of cepstral samples are determined from the quantile estimates and normalized to a constant range. A codebook decoding strategy is applied to determine the noisy models best matching the actual mixture of speech and noisy background. The proposed algorithms are evaluated side by side with conventional compensation schemes on connected Czech digits presented in various levels of background car noise. The resulting system provides an absolute word error rate (WER) reduction on 10-dB signal-to-noise ratio data of 8.7% and 37.7% for female neutral and LE speech, respectively, and of 8.7% and 32.8% for male neutral and LE speech, respectively, when compared to the baseline recognizer employing perceptual linear prediction (PLP) coefficients and cepstral mean and variance normalization.

72 citations

Patent
02 Oct 1995
TL;DR: In this paper, a translation word learning scheme for machine translation capable of learning translation words for each lexical rule separately and easily is proposed. But it does not address the problem of learning translations for each rule separately.
Abstract: A translation word learning scheme for a machine translation capable of learning translation words for each lexical rule separately and easily. In this scheme, a translation word for each original word is obtained by a machine translation using a translation dictionary storing headwords in the first language, a plurality of lexical rules for each headword, and at least one candidate translation word in the second language corresponding to each lexical rule. Then, a change of a translation word from that obtained by the machine translation to another translation word specified by a user is learned by registering a learning data indicating a headword, a top candidate translation word corresponding to a lexical rule applied in translating this headword, and the specified translation word. This specified translation word is used in subsequent translations only when an original word and a top candidate translation word for this original word obtained by the machine translation coincide with the headword and the top candidate translation word indicated in the learning data.

72 citations

Journal ArticleDOI
TL;DR: The strategy to present only the optimal result is not acceptable because it yields a substantial bias in error rate estimation, and alternative approaches for properly reporting classification accuracy are suggested.
Abstract: In biometric practice, researchers often apply a large number of different methods in a "trial-and-error" strategy to get as much as possible out of their data and, due to publication pressure or pressure from the consulting customer, present only the most favorable results. This strategy may induce a substantial optimistic bias in prediction error estimation, which is quantitatively assessed in the present manuscript. The focus of our work is on class prediction based on high-dimensional data (e.g. microarray data), since such analyses are particularly exposed to this kind of bias. In our study we consider a total of 124 variants of classifiers (possibly including variable selection or tuning steps) within a cross-validation evaluation scheme. The classifiers are applied to original and modified real microarray data sets, some of which are obtained by randomly permuting the class labels to mimic non-informative predictors while preserving their correlation structure. We assess the minimal misclassification rate over the different variants of classifiers in order to quantify the bias arising when the optimal classifier is selected a posteriori in a data-driven manner. The bias resulting from the parameter tuning (including gene selection parameters as a special case) and the bias resulting from the choice of the classification method are examined both separately and jointly. The median minimal error rate over the investigated classifiers was as low as 31% and 41% based on permuted uninformative predictors from studies on colon cancer and prostate cancer, respectively. We conclude that the strategy to present only the optimal result is not acceptable because it yields a substantial bias in error rate estimation, and suggest alternative approaches for properly reporting classification accuracy.

71 citations


Network Information
Related Topics (5)
Deep learning
79.8K papers, 2.1M citations
88% related
Feature extraction
111.8K papers, 2.1M citations
86% related
Convolutional neural network
74.7K papers, 2M citations
85% related
Artificial neural network
207K papers, 4.5M citations
84% related
Cluster analysis
146.5K papers, 2.9M citations
83% related
Performance
Metrics
No. of papers in the topic in previous years
YearPapers
2023271
2022562
2021640
2020643
2019633
2018528