scispace - formally typeset
D

Dong Yu

Researcher at Tencent

Publications -  389
Citations -  45733

Dong Yu is an academic researcher from Tencent. The author has contributed to research in topics: Artificial neural network & Word error rate. The author has an hindex of 72, co-authored 339 publications receiving 39098 citations. Previous affiliations of Dong Yu include Peking University & Microsoft.

Papers
More filters
Proceedings ArticleDOI

Context-dependent Deep Neural Networks for audio indexing of real-life data

TL;DR: It is found that for the best speaker-independent CD-DNN-HMM, with 32k senones trained on 2000h of data, the one-fourth reduction does carry over to inhomogeneous field data, and that DNN likelihood evaluation is a sizeable runtime factor even in the wide-beam context of generating rich lattices.
Proceedings ArticleDOI

Multiplex Word Embeddings for Selectional Preference Acquisition

TL;DR: This article proposed a multiplex word embedding model, which can be easily extended according to various relations among words, and the model can effectively distinguish words with respect to different relations without introducing unnecessary sparseness.
Proceedings ArticleDOI

Adaptive Permutation Invariant Training with Auxiliary Information for Monaural Multi-Talker Speech Recognition

TL;DR: This paper proposes to adapt the PIT models with auxiliary features such as pitch and i-vector, and to exploit the gender information with multi-task learning which jointly optimizes for the speech recognition and speaker-pair prediction.
Proceedings ArticleDOI

A long-contextual-span model of resonance dynamics for speech recognition: parameter learning and recognizer evaluation

TL;DR: A structured speech model that is equipped with the capability of jointly representing incomplete articulation and long-span co-articulation in natural human speech is presented, and the new model with rich parameter-free structure uses only the context-independent, single-modal Gaussian parameters.
Posted Content

Joint Separation and Denoising of Noisy Multi-talker Speech using Recurrent Neural Networks and Permutation Invariant Training

TL;DR: Deep bi-directional LSTM RNNs trained using uPIT in noisy environments can achieve large SDR and ESTOI improvements, when evaluated using known noise types, and that a single model is capable of handling multiple noise types with only a slight decrease in performance.