scispace - formally typeset
L

Lin-Shan Lee

Researcher at National Taiwan University

Publications -  407
Citations -  6885

Lin-Shan Lee is an academic researcher from National Taiwan University. The author has contributed to research in topics: Mandarin Chinese & Language model. The author has an hindex of 35, co-authored 405 publications receiving 6451 citations. Previous affiliations of Lin-Shan Lee include National Taiwan University of Science and Technology & Stanford University.

Papers
More filters

A distributed architecture for cooperative spoken dialogue agents with coherent dialogue state and history

TL;DR: Under this architecture, different spoken dialogue agents handling different domains can be developed independently, and cooperate with one another to respond to the user’s requests, while a user interface agent can access the correct spoken dialogue agent through a domain switching protocol, and carry over the dialogue state and history so as to keep the knowledge processed persistently and consistently across different domains.
Proceedings Article

Robust entropy-based endpoint detection for speech recognition in noisy environments.

TL;DR: This paper presents an entropy-based algorithm for accurate and robust endpoint detection for speech recognition under noisy environments that uses the spectral entropy to identify the speech segments accurately.
Proceedings ArticleDOI

A new framework for recognition of Mandarin syllables with tones using sub-syllabic units

TL;DR: A new structure for Mandarin syllable recognition is developed, in which the tones and base syllables are recognized jointly and a total of 574 subsyllabic unit models will be enough to provide improved recognition performance.
Proceedings ArticleDOI

A real-time Mandarin dictation machine for Chinese language with unlimited texts and very large vocabulary

TL;DR: A successfully implemented real-time Mandarin dictation machine which recognizes Mandarin speech with unlimited texts and very large vocabulary for the input of Chinese characters to computers is described.
Proceedings ArticleDOI

Audio Word2Vec: Unsupervised Learning of Audio Segment Representations Using Sequence-to-Sequence Autoencoder.

TL;DR: In this article, a parallel version of Word2Vec is proposed, which offers the vector representations of fixed dimensionality for variable-length audio segments, with very attractive real world applications such as query-by-example Spoken Term Detection (STD).