scispace - formally typeset
Open Access

Audiovisual speech processing in visual speech noise.

Jeesun Kim, +1 more
- pp 73-76
Reads0
Chats0
TLDR
Congruent AV speech facilitation and incongruent McGurk effects were tested by comparing percent correct syllable identification for full face visual speech stimuli compared to upper-face only conditions and results showed more accurate identification for congruent stimuli and less accurate responses for incongsruent ones.
Abstract
When the talker’s face (visual speech) can be seen, speech perception is both facilitated (for congruent visual speech) and interfered with (for incongruent visual speech). The current study investigated whether the degree of these visual speech effects was affected by the presence of an additional irrelevant talking face. In the experiment, auditory speech targets (vCv syllables) were presented in noise for subsequent speech identification. Participants were presented with the full display or upper-half (control) display of a talker’s face uttering single syllables either in central vision (Exp 1) or in the visual periphery (Exp 2). In addition, another talker was presented (silently uttering a sentence) either in the periphery (Exp 1) or in central vision (Exp 2). Participants’ eye-movements were monitored to ensure that participants always fixated centrally. Congruent AV speech facilitation and incongruent McGurk effects were tested by comparing percent correct syllable identification for full face visual speech stimuli compared to upper-face only conditions. The results showed more accurate identification for congruent stimuli and less accurate responses for incongruent ones (full face condition vs. the upper-half face control). The magnitude of the McGurk effect was greater when the face articulating the syllable was presented in central vision (with visual speech noise in the periphery) than when it was presented in the periphery (with central visual speech noise). The size of the congruent AV speech effect, however, did not differ as a function of central or peripheral presentation.

read more

Citations
More filters
Journal ArticleDOI

Audiovisual Binding for Speech Perception in Noise and in Aging

TL;DR: In this article, the authors exploit a specific paradigm in which a short audiovisual context made of coherent or incoherent speech material is displayed before an incongruent target likely to provide fusion.
Proceedings Article

Auditory speech processing is affected by visual speech in the periphery

TL;DR: Results show that peripheral visual speech affects speech recognition, and peripherally presented visual speech (full-face) facilitated identification of AV congruent stimuli compared to the upper-face control.
Proceedings ArticleDOI

Effect of context, rebinding and noise, on audiovisual speech fusion

TL;DR: The role of audiovisual coherence and noise in the binding process, in the framework of aud Giovisual speech scene analysis and the cocktail party effect, is evaluated.
Dissertation

Characterization of audiovisual binding and fusion in the framework of audiovisual speech scene analysis

TL;DR: It is shown that the AVSSA process enables to evaluate the coherence between auditory and visual features within a complex scene, in order to properly associate the adequate components of a given AV speech source, and provide to the fusion process an assessment of the AV coherence of the extracted source.

Modulating fusion in the McGurk effect by binding processes and contextual noise

TL;DR: The role of audiovisual coherence and noise in the binding process, in the framework of audiolabeled speech scene analysis and the cocktail party effect is concluded.
References
More filters
Journal ArticleDOI

Hearing lips and seeing voices

TL;DR: The study reported here demonstrates a previously unrecognised influence of vision upon speech perception, on being shown a film of a young woman's talking head in which repeated utterances of the syllable [ba] had been dubbed on to lip movements for [ga].
Journal ArticleDOI

Visual contribution to speech intelligibility in noise

TL;DR: In this article, the visual contribution to oral speech intelligibility was examined as a function of the speech-to-noise ratio and of the size of the vocabulary under test.
Journal ArticleDOI

Faces and text attract gaze independent of the task: Experimental data and computer model.

TL;DR: A well-known bottom-up computer model of saliency-driven attention that includes conspicuity maps for color, orientation, and intensity is refined by adding high-level semantic information and it is demonstrated that this significantly improves the ability to predict eye fixations in natural images.
Journal ArticleDOI

Assessing automaticity in audiovisual speech integration: evidence from the speeded classification task.

TL;DR: A syllabic version of the speeded classification paradigm, whereby response latencies to the first syllable of spoken word-like stimuli are slowed down when the second (irrelevant) syllable varies from trial to trial, suggested that audiovisual integration occurs prior to attentional selection in this paradigm.
Journal ArticleDOI

Gaze behavior in audiovisual speech perception: the influence of ocular fixations on the McGurk effect.

TL;DR: Findings demonstrate that the analysis of high spatial frequency information afforded by direct oral foveation is not necessary for the successful processing of visual speech information.