scispace - formally typeset
S

Shakti P. Rath

Researcher at Samsung

Publications -  32
Citations -  778

Shakti P. Rath is an academic researcher from Samsung. The author has contributed to research in topics: Acoustic model & Computer science. The author has an hindex of 9, co-authored 30 publications receiving 657 citations. Previous affiliations of Shakti P. Rath include Indian Institute of Technology Kanpur & University of Cambridge.

Papers
More filters
Proceedings ArticleDOI

Improved feature processing for Deep Neural Networks

TL;DR: The best result is obtained from splicing the baseline 40-dimensional speaker adapted features again across 9 frames, followed by reducing the dimension to 200 or 300 using another LDA, which is about 3% absolute better than the best GMM system.

Speech recognition and keyword spotting for low-resource languages : Babel project research at CUED

TL;DR: Using comparable systems over the five Option Period 1 languages indicates a strong correlation between ASR performance and KWS performance, and the approaches described show consistent trends over the languages investigated to date.
Proceedings ArticleDOI

Data augmentation for low resource languages

TL;DR: Two data augmentation schemes, semisupervised training and vocal tract length perturbation, are examined and combined on the Babel limited language pack configuration and consistent speech recognition performance gains can be obtained.
Proceedings ArticleDOI

Investigation of multilingual deep neural networks for spoken term detection

TL;DR: STT gains achieved through using multilingual bottleneck features in a Tandem configuration are shown to also apply to keyword search (KWS), and improvements in both STT and KWS were observed by incorporating language questions into the Tandem GMM-HMM decision trees for the training set languages.
Proceedings ArticleDOI

Language independent and unsupervised acoustic models for speech recognition and keyword spotting

TL;DR: This work considers a particular scenario where the target language is unseen in multi-language training and has limited language model training data, a limited lexicon, and acoustic training data without transcriptions, and unsupervised language dependent training yields performance gains.