Neural coding of continuous speech in auditory cortex during monaural and dichotic listening
Nai Ding,Jonathan Z. Simon +1 more
TLDR
These findings characterize how the spectrotemporal features of speech are encoded in human auditory cortex and establish a single-trial-based paradigm to study the neural basis underlying the cocktail party phenomenon.Abstract:
The cortical representation of the acoustic features of continuous speech is the foundation of speech perception. In this study, noninvasive magnetoencephalography (MEG) recordings are obtained from human subjects actively listening to spoken narratives, in both simple and cocktail party-like auditory scenes. By modeling how acoustic features of speech are encoded in ongoing MEG activity as a spectrotemporal response function, we demonstrate that the slow temporal modulations of speech in a broad spectral region are represented bilaterally in auditory cortex by a phase-locked temporal code. For speech presented monaurally to either ear, this phase-locked response is always more faithful in the right hemisphere, but with a shorter latency in the hemisphere contralateral to the stimulated ear. When different spoken narratives are presented to each ear simultaneously (dichotic listening), the resulting cortical neural activity precisely encodes the acoustic features of both of the spoken narratives, but slightly weakened and delayed compared with the monaural response. Critically, the early sensory response to the attended speech is considerably stronger than that to the unattended speech, demonstrating top-down attentional gain control. This attentional gain is substantial even during the subjects' very first exposure to the speech mixture and therefore largely independent of knowledge of the speech content. Together, these findings characterize how the spectrotemporal features of speech are encoded in human auditory cortex and establish a single-trial-based paradigm to study the neural basis underlying the cocktail party phenomenon.read more
Citations
More filters
Journal ArticleDOI
Cortical tracking of hierarchical linguistic structures in connected speech
Nai Ding,Lucia Melloni,Lucia Melloni,Lucia Melloni,Hang Zhang,Xing Tian,David Poeppel,David Poeppel +7 more
TL;DR: It is found that, during listening to connected speech, cortical activity of different timescales concurrently tracked the time course of abstract linguistic structures at different hierarchical levels, such as words, phrases and sentences.
Journal ArticleDOI
Mechanisms Underlying Selective Neuronal Tracking of Attended Speech at a “Cocktail Party”
Elana Zion Golumbic,Elana Zion Golumbic,Nai Ding,Stephan Bickel,Stephan Bickel,Peter Lakatos,Catherine A. Schevon,Guy M. McKhann,Robert R. Goodman,Ronald G. Emerson,Ashesh D. Mehta,Ashesh D. Mehta,Jonathan Z. Simon,David Poeppel,Charles E. Schroeder,Charles E. Schroeder +15 more
TL;DR: It is found that brain activity dynamically tracks speech streams using both low-frequency phase and high-frequency amplitude fluctuations and that optimal encoding likely combines the two.
Journal ArticleDOI
Emergence of neural encoding of auditory objects while listening to competing speakers
Nai Ding,Jonathan Z. Simon +1 more
TL;DR: Recording from subjects selectively listening to one of two competing speakers using magnetoencephalography indicates that concurrent auditory objects, even if spectrotemporally overlapping and not resolvable at the auditory periphery, are neurally encoded individually in auditory cortex and emerge as fundamental representational units for top-down attentional modulation and bottom-up neural adaptation.
Journal ArticleDOI
Attentional Selection in a Cocktail Party Environment Can Be Decoded from Single-Trial EEG
James O’Sullivan,Alan J. Power,Nima Mesgarani,Siddharth Rajaram,John J. Foxe,Barbara G. Shinn-Cunningham,Malcolm Slaney,Shihab A. Shamma,Edmund C. Lalor +8 more
TL;DR: It is shown that single-trial unaveraged EEG data can be decoded to determine attentional selection in a naturalistic multispeaker environment and a significant correlation between the EEG-based measure of attention and performance on a high-level attention task is shown.
Journal ArticleDOI
Speech rhythms and multiplexed oscillatory sensory coding in the human brain.
Joachim Gross,Nienke Hoogenboom,Gregor Thut,Philippe G. Schyns,Stefano Panzeri,Pascal Belin,Simon Garrod +6 more
TL;DR: A neuroimaging study reveals how coupled brain oscillations at different frequencies align with quasi-rhythmic features of continuous speech such as prosody, syllables, and phonemes.
References
More filters
Journal ArticleDOI
Gabor analysis of auditory midbrain receptive fields: spectro-temporal and binaural composition.
TL;DR: The properties of monaural STRFs and the relationship between ipsi- and contralateral inputs to neurons of the central nucleus of cat inferior colliculus (ICC) of cats are reported and it is shown that most interauralSTRF parameters are highly correlated bilaterally.
Journal ArticleDOI
Determination of activation areas in the human auditory cortex by means of synthetic aperture magnetometry
Anthony T. Herdman,Andreas Wollbrink,Andreas Wollbrink,Wilkin Chau,Ryouhei Ishii,Ryouhei Ishii,Bernhard Ross,Bernhard Ross,Christo Pantev,Christo Pantev +9 more
TL;DR: Investigating active cortical areas associated with magnetically recorded transient and steady-state auditory evoked responses suggests that SAM is a useful technique for imaging cortical structures involved in processing perceptual information.
Journal ArticleDOI
Attention-driven auditory cortex short-term plasticity helps segregate relevant sounds from noise.
Jyrki Ahveninen,Matti Hämäläinen,Matti Hämäläinen,Iiro P. Jääskeläinen,Seppo P. Ahlfors,Seppo P. Ahlfors,Samantha Huang,Fa-Hsuan Lin,Fa-Hsuan Lin,Tommi Raij,Mikko Sams,Christos E. Vasios,John W. Belliveau,John W. Belliveau +13 more
TL;DR: A simple gain model alone cannot explain auditory selective attention, and attention-driven short-term plasticity retunes neurons to segregate relevant sounds from noise in noisy environments.
Journal ArticleDOI
Neuromagnetic responses to frequency-tagged sounds: A new method to follow inputs from each ear to the human auditory cortex during binaural hearing
TL;DR: A novel method is introduced that allows, for the first time, to selectively follow these inputs in humans up to the cortex during binaural hearing, using neuromagnetic cortical responses to amplitude-modulated continuous tones, with different modulation frequencies at each ear.
Journal ArticleDOI
The neural processing of masked speech: evidence for different mechanisms in the left and right temporal lobes.
TL;DR: Functional imaging results reveal that masking speech with speech leads to bilateral superior temporal gyrus (STG) activation relative to a speech-in-noise baseline, while masking Speech and two additional maskers derived from the original speech were investigated, showing that masks can arise through two parallel neural systems, in the left and right temporal lobes.
Related Papers (5)
Emergence of neural encoding of auditory objects while listening to competing speakers
Nai Ding,Jonathan Z. Simon +1 more