scispace - formally typeset
Search or ask a question
Topic

Viseme

About: Viseme is a research topic. Over the lifetime, 865 publications have been published within this topic receiving 17889 citations.


Papers
More filters
Journal ArticleDOI
TL;DR: An approach to ALR is proposed that acknowledges that this information is missing but assumes that it is substituted or deleted in a systematic way that can be modelled, and a system that learns such a model and then incorporates it into decoding, which is realised as a cascade of weighted finite-state transducers.

32 citations

Proceedings ArticleDOI
25 Mar 2012
TL;DR: A multiview dataset using connected words that can be analysed by an automatic system, based on linear predictive trackers and active appearance models, and human lip-readers, and the automatic system is good at guessing its fallibility.
Abstract: Computer lip-reading is one of the great signal processing challenges. Not only is the signal noisy, it is variable. However it is almost unknown to compare the performance with human lip-readers. Partly this is because of the paucity of human lip-readers and partly because most automatic systems only handle data that are trivial and therefore not representative of human speech. Here we generate a multiview dataset using connected words that can be analysed by an automatic system, based on linear predictive trackers and active appearance models, and human lip-readers. The automatic system we devise has a viseme accuracy of ≈ 46% which is comparable to poor professional human lip-readers. However, unlike human lip-readers our system is good at guessing its fallibility.

32 citations

Journal ArticleDOI
TL;DR: A multiple inputs-driven realistic facial animation system based on 3-D virtual head for human-machine interface based on parameterized model and muscular model is proposed, and the objective and subjective experiments show that the system is suitable forhuman-machine interaction.
Abstract: A multiple inputs-driven realistic facial animation system based on 3-D virtual head for human–machine interface is proposed. The system can be driven independently by video, text, and speech, thus can interact with humans through diverse interfaces. The combination of parameterized model and muscular model is used to obtain a tradeoff between computational efficiency and high realism of 3-D facial animation. The online appearance model is used to track 3-D facial motion from video in the framework of particle filtering, and multiple measurements, i.e., pixel color value of input image and Gabor wavelet coefficient of illumination ratio image, are infused to reduce the influence of lighting and person dependence for the construction of online appearance model. The tri-phone model is used to reduce the computational consumption of visual co-articulation in speech synchronized viseme synthesis without sacrificing any performance. The objective and subjective experiments show that the system is suitable for human–machine interaction.

32 citations

Proceedings Article
01 Jan 2002
TL;DR: Several extensions to the original coarticulation algorithm of Cohen and Massaro are implemented, including an optimization to improve performance as well as special treatment of closure and release phase of bilabial stops and other phonemes.
Abstract: We present a method for generating realistic speech-synchronized facial animations using a physicsbased approach and support for coarticulation, i.e. the coloring of a speech segment by surrounding segments. We have implemented several extensions to the original coarticulation algorithm of Cohen and Massaro [Cohen93]. The enhancements include an optimization to improve performance as well as special treatment of closure and release phase of bilabial stops and other phonemes. Furthermore, for phonemes that are shorter than the sampling intervals of the algorithm and might therefore be missed, additional key frames are created to ensure their impact onto the animation.

32 citations


Network Information
Related Topics (5)
Vocabulary
44.6K papers, 941.5K citations
78% related
Feature vector
48.8K papers, 954.4K citations
76% related
Feature extraction
111.8K papers, 2.1M citations
75% related
Feature (computer vision)
128.2K papers, 1.7M citations
74% related
Unsupervised learning
22.7K papers, 1M citations
73% related
Performance
Metrics
No. of papers in the topic in previous years
YearPapers
20237
202212
202113
202039
201919
201822