Improved Bottleneck Features Using Pretrained Deep Neural Networks.
Citations
2,817 citations
701 citations
464 citations
Cites background or methods from "Improved Bottleneck Features Using ..."
...The work in [10] shows that DNN trained bottleneck feature reduces word error rate by 16% relatively on a large vocabulary business search task...
[...]
...Besides CD-DNN-HMMs, DNN can also be used to provide the bottle-neck feature vectors of the GMM in a GMM-HMM system [10][11] Both applications of DNN in ASR achieved significant accuracy improvement....
[...]
408 citations
360 citations
Cites background or methods from "Improved Bottleneck Features Using ..."
...However, in many other areas such as speech and auto-encoder, fully connected neural network is also a major type of workload, such as networks presented on work [39][40][41][42][43][44]....
[...]
...Figure 17 presents a design space of a bottleneck network, which is also frequently used in prior work[40][41][42]....
[...]
References
16,717 citations
5,150 citations
"Improved Bottleneck Features Using ..." refers methods in this paper
...Because 〈 h 〉03*45 is extremely expensive to compute exactly, the contrastive divergence (CD) approximation to the gradient is used, where 〈 h 〉03*45 is replaced by running the Gibbs sampler initialized at the data for one full step [12]....
[...]
3,120 citations
"Improved Bottleneck Features Using ..." refers methods in this paper
...For example, the DNN-HMM which exploits the discriminative learning ability of pretrained DNNs and the sequential modeling ability of hidden Markov models (HMMs) outperformed the conventional Gaussian mixture model (GMM)-HMM for both phoneme recognition [6][7] and large vocabulary speech recognition [ 8 ] tasks.,We are also interested in knowing how bottleneck features perform compared to the context-dependent DNN-HMMs developed recently [ 8 ] For these purposes, we conducted a series of experiments using the Windows Live Search for Mobile (WLS4M) corpus collected from real users of a smartphone application for business search [13][14].,The lexicon and trigram language model (LM) used for decoding were the same as used in our previous work [ 8 ].,In all the results reported here, we followed the DNN training recipe in [ 8 ]....
[...]
2,036 citations
1,767 citations
"Improved Bottleneck Features Using ..." refers methods in this paper
...For example, the DNN-HMM which exploits the discriminative learning ability of pretrained DNNs and the sequential modeling ability of hidden Markov models (HMMs) outperformed the conventional Gaussian mixture model (GMM)-HMM for both phoneme recognition [ 6 ][7] and large vocabulary speech recognition [8] tasks....
[...]