scispace - formally typeset
Proceedings ArticleDOI

Lip signatures for automatic person recognition

TLDR
In this article, the authors evaluated lip features for person recognition and compared the performance with that of the acoustic signal, and found that the performance of the upper lip was considerably better than the lower lip, achieving 15% and 35% identification error rates respectively.
Abstract
This paper evaluates lip features for person recognition, and compares the performance with that of the acoustic signal. Recognition accuracy is found to be equivalent in the two domains, agreeing with the findings of Chibelushi (1997). The optimum dynamic window length for both acoustic and visual modalities is found to be about 100 ms. Recognition performance of the upper lip is considerably better than the lower lip, achieving 15% and 35% identification error rates respectively, using a single digit test and training token.

read more

Citations
More filters
Journal ArticleDOI

Behavioural biometrics: a survey and classification

TL;DR: A survey and classification of the state-of-the-art in behavioural biometrics which is based on skills, style, preference, knowledge, motor-skills or strategy used by people while accomplishing different everyday tasks.
Journal ArticleDOI

A review of speech-based bimodal recognition

TL;DR: The components of bimodal recognizers are reviewed, the accuracy of bIModal recognition is discussed, some outstanding research issues as well as possible application domains are highlighted, and the combination of auditory and visual modalities promises higher recognition accuracy and robustness than can be obtained with a single modality.
Journal ArticleDOI

Analysis of lip geometric features for audio-visual speech recognition

TL;DR: Experimental results show that among the geometric visual features analyzed, lip vertical aperture is the most relevant; and the visual feature vector formed by vertical and horizontal lip apertures and the first-order derivative of the lip corner angle leads to the best recognition results.
Journal ArticleDOI

Local Ordinal Contrast Pattern Histograms for Spatiotemporal, Lip-Based Speaker Authentication

TL;DR: A texture descriptor called local ordinal contrast pattern with a dynamic texture representation called three orthogonal planes is used to represent both the appearance and dynamics features observed in visual speech to suggest that there is enough discriminative information in the mouth-region to enable its use as a primary biometric trait.
Book ChapterDOI

Taxonomy of Behavioural Biometrics

TL;DR: This chapter presents a taxonomy of the latest behavioural biometrics, including some future oriented approaches, and addresses privacy issues which arise or might arise in the future with the use of behavioural biometric approaches.
References
More filters
Journal ArticleDOI

Human and machine recognition of faces: a survey

TL;DR: A critical survey of existing literature on human and machine recognition of faces is presented, followed by a brief overview of the literature on face recognition in the psychophysics community and a detailed overview of move than 20 years of research done in the engineering community.
Journal ArticleDOI

Feature extraction from faces using deformable templates

TL;DR: In this article, a deformable template is used to detect and describe features of faces using deformable templates and an energy function is defined which links edges, peaks, and valleys in the image intensity to corresponding properties of the template.

Cepstrum analysis technique for automatic speaker verification

S. Furui
TL;DR: New techniques for automatic speaker verification using telephone speech based on a set of functions of time obtained from acoustic analysis of a fixed, sentence-long utterance using a new time warping method using a dynamic programming technique.
Journal ArticleDOI

Cepstral analysis technique for automatic speaker verification

TL;DR: In this paper, a set of functions of time obtained from acoustic analysis of a fixed, sentence-long utterance are extracted by means of LPC analysis successively throughout an utterance to form time functions, and frequency response distortions introduced by transmission systems are removed.
Proceedings ArticleDOI

"Eigenlips" for robust speech recognition

TL;DR: This study improves the performance of a hybrid connectionist speech recognition system by incorporating visual information about the corresponding lip movements by using a new visual front end, and an alternative architecture for combining the visual and acoustic information.