scispace - formally typeset
D

DeLiang Wang

Researcher at Ohio State University

Publications -  475
Citations -  28623

DeLiang Wang is an academic researcher from Ohio State University. The author has contributed to research in topics: Speech processing & Speech enhancement. The author has an hindex of 82, co-authored 440 publications receiving 23687 citations. Previous affiliations of DeLiang Wang include Massachusetts Institute of Technology & Tsinghua University.

Papers
More filters
Proceedings ArticleDOI

Attentive Training: A New Training Framework for Talker-independent Speaker Extraction

TL;DR: In this article , a deep neural network is trained to create a representation for the first speaker and utilize it to extract or track that speaker from a multitalker noisy mixture, and the results demonstrate the superiority of attentive training over widely used permutation invariant training for talker-independent speaker extraction.
Journal ArticleDOI

Fusing Bone-Conduction and Air-Conduction Sensors for Complex-Domain Speech Enhancement

TL;DR: This study proposes an attention-based fusion method to combine the strengths of AC and BC signals and perform complex spectral mapping for speech enhancement, and shows that the sensor fusion method is superior to single-sensor counterparts, especially in low SNR conditions.
Journal ArticleDOI

KalmanNet: A Learnable Kalman Filter for Acoustic Echo Cancellation

TL;DR: In this article , the authors integrate the frequency domain Kalman filter and deep neural networks (DNNs) into a hybrid method, called KalmanNet, to leverage the advantages of deep learning and adaptive filtering algorithms.
Book

Neural Information Processing: 13th International Conference, ICONIP 2006, Hong Kong, China, October 3-6, 2006, Proceedings, Part II (Lecture Notes in Computer Science)

TL;DR: This presentation presented at the 13th International Conference of the ICONIP on Neural Information Processing focused on the development of neural information processing models for discrete-time decision-making.
Posted Content

Recurrent Deep Stacking Networks for Speech Recognition.

TL;DR: A more efficient yet comparable substitute to RDSN, Bi- Pass Stacking Network (BPSN) is proposed, to add phoneme-level information into acoustic models, transforming anoustic model to the combination of an acoustic model and a phonemic-level N-gram model.