scispace - formally typeset
W

Wei Xia

Researcher at University of Texas at Dallas

Publications -  16
Citations -  239

Wei Xia is an academic researcher from University of Texas at Dallas. The author has contributed to research in topics: Computer science & Speaker recognition. The author has an hindex of 6, co-authored 12 publications receiving 171 citations.

Papers
More filters
Proceedings ArticleDOI

UTD-CRSS Systems for 2018 NIST Speaker Recognition Evaluation

TL;DR: This study presents systems submitted by the Center for Robust Speech Systems from UTDallas to NIST SRE 2018, and investigates three alternative front-end speaker embedding frameworks, finding them to be both complementary and effective in achieving overall improved speaker recognition performance.
Proceedings ArticleDOI

Cross-lingual Text-independent Speaker Verification Using Unsupervised Adversarial Discriminative Domain Adaptation

TL;DR: Data analysis of ADDA adapted speaker embedding shows that the learned speaker embeddings can perform well on speaker classification for the target domain data, and are less dependent with respect to the shift in language.
Posted Content

Self-supervised Text-independent Speaker Verification using Prototypical Momentum Contrastive Learning

TL;DR: A simple contrastive learning approach (SimCLR) with a momentum contrastive (MoCo) learning framework, where the MoCo speaker embedding system utilizes a queue to maintain a large set of negative examples, is examined.
Proceedings ArticleDOI

Automated evaluation of non-native English pronunciation quality: combining knowledge- and data-driven features at multiple time scales.

TL;DR: This paper experiments with both data-driven and knowledge-inspired features that capture degree of nativeness from pauses in speech, speaking rate, rhythm/stress, and goodness of phone pronunciation, and finds that highly accurate automated assessment can be attained using a small diverse set of intuitive and interpretable features.
Proceedings ArticleDOI

A dynamic model for behavioral analysis of couple interactions using acoustic features.

TL;DR: It is shown that dynamic models can achieve up to 10% relative improvement, compared to static models, which suggests that the human behavioral interaction is a non-linear process, and the resulting latent-state labels may provide new insights to domain experts.