scispace - formally typeset
Open AccessJournal ArticleDOI

Neural coding of continuous speech in auditory cortex during monaural and dichotic listening

Nai Ding, +1 more
- 01 Jan 2012 - 
- Vol. 107, Iss: 1, pp 78-89
TLDR
These findings characterize how the spectrotemporal features of speech are encoded in human auditory cortex and establish a single-trial-based paradigm to study the neural basis underlying the cocktail party phenomenon.
Abstract
The cortical representation of the acoustic features of continuous speech is the foundation of speech perception. In this study, noninvasive magnetoencephalography (MEG) recordings are obtained from human subjects actively listening to spoken narratives, in both simple and cocktail party-like auditory scenes. By modeling how acoustic features of speech are encoded in ongoing MEG activity as a spectrotemporal response function, we demonstrate that the slow temporal modulations of speech in a broad spectral region are represented bilaterally in auditory cortex by a phase-locked temporal code. For speech presented monaurally to either ear, this phase-locked response is always more faithful in the right hemisphere, but with a shorter latency in the hemisphere contralateral to the stimulated ear. When different spoken narratives are presented to each ear simultaneously (dichotic listening), the resulting cortical neural activity precisely encodes the acoustic features of both of the spoken narratives, but slightly weakened and delayed compared with the monaural response. Critically, the early sensory response to the attended speech is considerably stronger than that to the unattended speech, demonstrating top-down attentional gain control. This attentional gain is substantial even during the subjects' very first exposure to the speech mixture and therefore largely independent of knowledge of the speech content. Together, these findings characterize how the spectrotemporal features of speech are encoded in human auditory cortex and establish a single-trial-based paradigm to study the neural basis underlying the cocktail party phenomenon.

read more

Content maybe subject to copyright    Report

Citations
More filters
Posted Content

Inference of the Selective Auditory Attention using Sequential LMMSE Estimation

TL;DR: In this article, the authors presented a novel framework to decode the attention of a listener within trial durations of the order of two seconds, which comprises of three modules: 1) Dynamic estimation of the temporal response functions (TRF) in every trial using a sequential linear minimum mean squared error (LMMSE) estimator, 2) Extract the N1-P2 peak of the estimated TRF that serves as a marker related to the attentional state and 3) Obtain a probabilistic measure of the attentiveal state using a support vector machine followed by a
Journal Article

Real time fpga implementation for attentional selection in multi-speaker environment

TL;DR: This novel framework bridges the gap between the most recent advancements in speech processing technologies and speech prosthesis research and moves us closer to the development of cognitively controlled hearable devices for the hearing impaired.
Book ChapterDOI

Interfacing Hearing Implants with the Brain: Closing the Loop with Intracochlear Brain Recordings

TL;DR: An overview of the ongoing research towards neuro-steered hearing implants is presented and it is shown that Integrating a brain-computer interface (BCI) can alleviate the issues with the current fitting paradigms.

A Causal Role of Human Laryngeal Motor Cortex in 1 Speech Lexical Tone and Voicing Perception

TL;DR: Results reveal a causal role of the LMC in speech perception that it mirrors its distinctive functions in controlling pitch and voicing during production and it is upregulated and shifts from a left-lateralized to a bilateral engagement in adverse listening conditions as a means of compensation.
Proceedings ArticleDOI

Esaa: An Eeg-Speech Auditory Attention Detection Database

TL;DR: The EEG-Speech Auditory Attention Detection (AAD) database as mentioned in this paper contains 12.7 hours of data collected from 20 subjects and achieved an accuracy of 84.6% and 84.3% for speaker and speaker locus attention detection with 64-channel and 1-second decision window, respectively.
References
More filters
Book

Elements of information theory

TL;DR: The author examines the role of entropy, inequality, and randomness in the design of codes and the construction of codes in the rapidly changing environment.
Journal ArticleDOI

The cortical organization of speech processing

TL;DR: A dual-stream model of speech processing is outlined that assumes that the ventral stream is largely bilaterally organized — although there are important computational differences between the left- and right-hemisphere systems — and that the dorsal stream is strongly left- Hemisphere dominant.
Journal ArticleDOI

Some Experiments on the Recognition of Speech, with One and with Two Ears

TL;DR: In this paper, the relation between the messages received by the two ears was investigated, and two types of test were reported: (a) the behavior of a listener when presented with two speech signals simultaneously (statistical filtering problem) and (b) behavior when different speech signals are presented to his two ears.
Journal ArticleDOI

Speech recognition with primarily temporal cues.

TL;DR: Nearly perfect speech recognition was observed under conditions of greatly reduced spectral information; the presentation of a dynamic temporal pattern in only a few broad spectral regions is sufficient for the recognition of speech.
Journal ArticleDOI

Electrical Signs of Selective Attention in the Human Brain

TL;DR: Auditory evoked potentials were recorded from the vertex of subjects who listened selectively to a series of tone pipping in one ear and ignored concurrent tone pips in the other ear to study the response set established to recognize infrequent, higher pitched tone pipped in the attended series.
Related Papers (5)