scispace - formally typeset
D

Dong Yu

Researcher at Tencent

Publications -  389
Citations -  45733

Dong Yu is an academic researcher from Tencent. The author has contributed to research in topics: Artificial neural network & Word error rate. The author has an hindex of 72, co-authored 339 publications receiving 39098 citations. Previous affiliations of Dong Yu include Peking University & Microsoft.

Papers
More filters
Journal ArticleDOI

An introduction to voice search

TL;DR: This article categorized spoken dialog technology into form filling, call routing, and voice search, and reviewed the voice search technology.
Journal ArticleDOI

Structured speech modeling

TL;DR: This paper shows how the use of resonance target parameters and their temporal filtering enables joint modeling of long-span coarticulation and phonetic reduction effects and demonstrates superior recognizer performance over a modern hidden Markov model-based system.
Proceedings ArticleDOI

On parallelizability of stochastic gradient descent for speech DNNS

TL;DR: It is shown that data-parallel training efficiency can be improved by increasing the minibatch size (through a combination of AdaGrad and automatic adjustments of learning rate and minibATCH size) and data compression, and that model parallelism is optimal with only 3 GPUs in a single server, while data parallelism with a minibatches size of 1024 does not even scale to 2 GPUs.
Book

Foundations and Trends in Signal Processing: DEEP LEARNING – Methods and Applications

Li Deng, +1 more
TL;DR: Deep Learning: Methods and Applications as mentioned in this paper provides an overview of general deep learning methodology and its applications to a variety of signal and information processing tasks, including natural language and text processing, information retrieval, and multimodal information processing empowered by multitask deep learning.
Proceedings ArticleDOI

Pipelined Back-Propagation for Context-Dependent Deep Neural Networks.

TL;DR: It is shown that the pipelined approximation to BP, which parallelizes computation with respect to layers, is an efficient way of utilizing multiple GPGPU cards in a single server.