scispace - formally typeset
H

Hao Zhang

Researcher at Carnegie Mellon University

Publications -  6
Citations -  323

Hao Zhang is an academic researcher from Carnegie Mellon University. The author has contributed to research in topics: Speaker recognition & Artificial neural network. The author has an hindex of 5, co-authored 6 publications receiving 305 citations.

Papers
More filters
Journal ArticleDOI

Speaker adaptive training of deep neural network acoustic models using i-vectors

TL;DR: This paper ports the idea of SAT to deep neural networks (DNNs), and proposes a framework to perform feature-space SAT for DNNs, using i-vectors as speaker representations and an adaptation neural network to derive speaker-normalized features.
Proceedings ArticleDOI

Towards Speaker Adaptive Training of Deep Neural Network Acoustic Models

TL;DR: Experiments show that compared with the baseline DNN, the SAT-DNN model brings 7.5% and 6.0% relative improvement when DNN inputs are speaker-independent and speakeradapted features respectively.
Proceedings ArticleDOI

Improvements to speaker adaptive training of deep neural networks

TL;DR: Different methods to further improve and extend SAT-DNN to improve tasks including bottleneck feature (BNF) generation, convolutional neural network (CNN) acoustic modeling and multilingual DNN-based feature extraction are presented.
Proceedings ArticleDOI

Distributed learning of multilingual DNN feature extractors using GPUs.

TL;DR: This paper proposes the DistModel and DistLang frameworks which distribute feature extractor learning by models and languages respectively and investigates strategies to accelerate the learning process over multiple GPU cards.
Proceedings ArticleDOI

Semi-supervised training in low-resource ASR and KWS

TL;DR: A set of experiments on low-resource languages in telephony speech quality in Assamese, Bengali, Lao, Haitian, Zulu, and Tamil are presented, demonstrating the impact that semi-supervised training and speaker adaptation techniques can have, in particular learning robust bottle-neck features on the test data.