scispace - formally typeset
H

H. S. Jayanna

Researcher at Siddaganga Institute of Technology

Publications -  44
Citations -  319

H. S. Jayanna is an academic researcher from Siddaganga Institute of Technology. The author has contributed to research in topics: Speaker recognition & Mel-frequency cepstrum. The author has an hindex of 9, co-authored 39 publications receiving 238 citations. Previous affiliations of H. S. Jayanna include Indian Institute of Technology Guwahati & Jain University.

Papers
More filters
Journal ArticleDOI

Analysis, Feature Extraction, Modeling and Testing Techniques for Speaker Recognition

TL;DR: An overview of the major techniques developed in each of these stages of speaker recognition system is given, along with a discussion on the possible future direction for the development of techniques in each stage.
Journal ArticleDOI

Multiple frame size and rate analysis for speaker recognition under limited data condition

TL;DR: This work demonstrates the usefulness of multiple frame size and rate (MFSR) analysis for speaker recognition under limited data condition and demonstrates that the MFSR analysis outperforms the Gaussian mixture model-universal background model (GMM-UBM) performance, the most widely used modelling technique.
Journal ArticleDOI

A spoken query system for the agricultural commodity prices and weather information access in Kannada language

TL;DR: A spoken query system is demonstrated which can be used to access the latest agricultural commodity prices and weather information in Kannada language using mobile phone.
Proceedings ArticleDOI

Fuzzy Vector Quantization for speaker recognition under limited data conditions

TL;DR: Among these FVQ shows significant improved performance compared to DTM and CVQ, and this work performs an experimental evaluation of three simple modelling techniques namely, direct template matching (DTM), crisp vector quantization (CVQ) and fuzzy vectorquantization (FVQ).
Journal ArticleDOI

An experimental comparison of modelling techniques for speaker recognition under limited data condition

TL;DR: It is proposed that the combined LVQ and GMM-UBM classifier provides relatively better performance compared to all the individual as well as combined classifiers.