The Right Temporoparietal Junction Supports Speech Tracking During Selective Listening: Evidence from Concurrent EEG-fMRI

Listening selectively to one out of several competing speakers in a “cocktail party” situation is a highly demanding task. It relies on a widespread cortical network, including auditory sensory, but also frontal and parietal brain regions involved in controlling auditory attention. Previous work has shown that, during selective listening, ongoing neural activity in auditory sensory areas is dominated by the attended speech stream, whereas competing input is suppressed. The relationship between these attentional modulations in the sensory tracking of the attended speech stream and frontoparietal activity during selective listening is, however, not understood. We studied this question in young, healthy human participants (both sexes) using concurrent EEG-fMRI and a sustained selective listening task, in which one out of two competing speech streams had to be attended selectively. An EEG-based speech envelope reconstruction method was applied to assess the strength of the cortical tracking of the to-be-attended and the to-be-ignored stream during selective listening. Our results show that individual speech envelope reconstruction accuracies obtained for the to-be-attended speech stream were positively correlated with the amplitude of sustained BOLD responses in the right temporoparietal junction, a core region of the ventral attention network. This brain region further showed task-related functional connectivity to secondary auditory cortex and regions of the frontoparietal attention network, including the intraparietal sulcus and the inferior frontal gyrus. This suggests that the right temporoparietal junction is involved in controlling attention during selective listening, allowing for a better cortical tracking of the attended speech stream. SIGNIFICANCE STATEMENT Listening selectively to one out of several simultaneously talking speakers in a “cocktail party” situation is a highly demanding task. It activates a widespread network of auditory sensory and hierarchically higher frontoparietal brain regions. However, how these different processing levels interact during selective listening is not understood. Here, we investigated this question using fMRI and concurrently acquired scalp EEG. We found that activation levels in the right temporoparietal junction correlate with the sensory representation of a selectively attended speech stream. In addition, this region showed significant functional connectivity to both auditory sensory and other frontoparietal brain areas during selective listening. This suggests that the right temporoparietal junction contributes to controlling selective auditory attention in “cocktail party” situations.

[1]  Adrian K. C. Lee,et al.  Using neuroimaging to understand the cortical mechanisms of auditory selective attention , 2014, Hearing Research.

[2]  John J. Foxe,et al.  Auditory selective attention and processing in children with attention-deficit/hyperactivity disorder , 2012, Clinical Neurophysiology.

[3]  T. Lunner,et al.  The Ease of Language Understanding (ELU) model: theoretical, empirical, and clinical advances , 2013, Front. Syst. Neurosci..

[4]  Lee M. Miller,et al.  Auditory attentional control and selection during cocktail party listening. , 2010, Cerebral cortex.

[5]  Jonathan Z. Simon,et al.  Adaptive Temporal Encoding Leads to a Background-Insensitive Cortical Representation of Speech , 2013, The Journal of Neuroscience.

[6]  J. Simon,et al.  Emergence of neural encoding of auditory objects while listening to competing speakers , 2012, Proceedings of the National Academy of Sciences.

[7]  A. Bronkhorst The cocktail-party problem revisited: early processing and selection of multi-talker speech , 2015, Attention, Perception, & Psychophysics.

[8]  O. Bertrand,et al.  Effects of Selective Attention on the Electrophysiological Representation of Concurrent Sounds in the Human Auditory Cortex , 2007, The Journal of Neuroscience.

[9]  D. Poeppel,et al.  Mechanisms Underlying Selective Neuronal Tracking of Attended Speech at a “Cocktail Party” , 2013, Neuron.

[10]  John J. Foxe,et al.  Attentional Selection in a Cocktail Party Environment Can Be Decoded from Single-Trial EEG. , 2015, Cerebral cortex.

[11]  Matti S. Hämäläinen,et al.  Lateralized parietotemporal oscillatory phase synchronization during auditory selective attention , 2014, NeuroImage.

[12]  Antoine J. Shahin,et al.  Attentional Gain Control of Ongoing Cortical Speech Representations in a “Cocktail Party” , 2010, The Journal of Neuroscience.

[13]  N. Mesgarani,et al.  Selective cortical representation of attended speaker in multi-talker speech perception , 2012, Nature.

[14]  Larry J. Seidman,et al.  Distinct cortical networks activated by auditory attention and working memory load , 2013, NeuroImage.

[15]  Gavin M. Bidelman,et al.  Bilinguals at the “cocktail party”: Dissociable neural activity in auditory–linguistic brain regions reveals neurobiological basis for nonnative listeners’ speech-in-noise recognition deficits , 2015, Brain and Language.

[16]  Susan L. Whitfield-Gabrieli,et al.  Conn: A Functional Connectivity Toolbox for Correlated and Anticorrelated Brain Networks , 2012, Brain Connect..

[17]  Michael J. Crosse,et al.  Eye Can Hear Clearly Now: Inverse Effectiveness in Natural Audiovisual Speech Processing Relies on Long-Term Crossmodal Temporal Integration , 2016, The Journal of Neuroscience.

[18]  Riitta Hari,et al.  Left Superior Temporal Gyrus Is Coupled to Attended Speech in a Cocktail-Party Auditory Scene , 2016, The Journal of Neuroscience.

[19]  Mounya Elhilali,et al.  Competing Streams at the Cocktail Party: Exploring the Mechanisms of Attention and Temporal Integration , 2010, The Journal of Neuroscience.

[20]  Arnaud Delorme,et al.  EEGLAB: an open source toolbox for analysis of single-trial EEG dynamics including independent component analysis , 2004, Journal of Neuroscience Methods.

[21]  E. C. Cmm,et al.  on the Recognition of Speech, with , 2008 .

[22]  Josh H. McDermott The cocktail party problem , 2009, Current Biology.

[23]  Maarten De Vos,et al.  Decoding the attended speech stream with multi-channel EEG: implications for online, daily-life applications , 2015, Journal of neural engineering.

[24]  Eric Larson,et al.  Switching auditory attention using spatial and non-spatial features recruits different cortical networks , 2014, NeuroImage.

[25]  M. Corbetta,et al.  The Reorienting System of the Human Brain: From Environment to Theory of Mind , 2008, Neuron.

[26]  Barak A. Pearlmutter,et al.  The VESPA: A method for the rapid estimation of a visual evoked potential , 2006, NeuroImage.

[27]  Jyrki Ahveninen,et al.  Brain Networks of Novelty-Driven Involuntary and Cued Voluntary Auditory Attention Shifting , 2012, PloS one.

[28]  Edmund C. Lalor,et al.  The Multivariate Temporal Response Function (mTRF) Toolbox: A MATLAB Toolbox for Relating Neural Signals to Continuous Stimuli , 2016, Front. Hum. Neurosci..

[29]  Barbara G Shinn-Cunningham,et al.  Cerebral Cortex doi:10.1093/cercor/bhs359 Auditory Spatial Attention Representations in the Human Cerebral Cortex , 2012 .

[30]  Alessandro Presacco,et al.  Effect of informational content of noise on speech representation in the aging midbrain and cortex. , 2016, Journal of neurophysiology.

[31]  Robert J. Zatorre,et al.  Speech-in-noise perception in musicians: A review , 2017, Hearing Research.

[32]  Hongkeun Kim Involvement of the dorsal and ventral attention networks in oddball stimulus processing: A meta‐analysis , 2014, Human brain mapping.

[33]  Nina Kraus,et al.  Sensory-cognitive interaction in the neural encoding of speech in noise: a review. , 2010, Journal of the American Academy of Audiology.

[34]  Rami K. Niazy,et al.  Removal of FMRI environment artifacts from EEG data using optimal basis sets , 2005, NeuroImage.

[35]  Sebastian Puschmann,et al.  Mapping the spatiotemporal dynamics of processing task‐relevant and task‐irrelevant sound feature changes using concurrent EEG‐fMRI , 2016, Human brain mapping.

[36]  Matti S. Hämäläinen,et al.  Dynamic Oscillatory Processes Governing Cued Orienting and Allocation of Auditory Attention , 2013, Journal of Cognitive Neuroscience.

[37]  R. Näätänen,et al.  Selection of speech messages in free-field listening. , 1993, Neuroreport.

[38]  Barbara G. Shinn-Cunningham,et al.  Short-Term Memory for Space and Time Flexibly Recruit Complementary Sensory-Biased Frontal Lobe Attention Networks , 2015, Neuron.

[39]  Stefan Debener,et al.  Target Speaker Detection with Concealed EEG Around the Ear , 2016, Front. Neurosci..

[40]  Jonathan Z Simon,et al.  The encoding of auditory objects in auditory cortex: insights from magnetoencephalography. , 2015, International journal of psychophysiology : official journal of the International Organization of Psychophysiology.

[41]  Sach Aj Auditory spatial attention. , 2000 .

[42]  Edmund C Lalor,et al.  Endogenous auditory spatial attention modulates obligatory sensory activity in auditory cortex. , 2011, Cerebral cortex.

[43]  Mikko Sams,et al.  Attention-driven auditory cortex short-term plasticity helps segregate relevant sounds from noise , 2011, Proceedings of the National Academy of Sciences.

[44]  J. Rauschecker,et al.  Are you listening? Brain activation associated with sustained nonspatial auditory attention in the presence and absence of stimulation , 2014, Human brain mapping.

[45]  I. Winkler,et al.  MMN and attention: competition for deviance detection. , 2003, Psychophysiology.

[46]  J. Simon,et al.  Neural coding of continuous speech in auditory cortex during monaural and dichotic listening. , 2012, Journal of neurophysiology.

[47]  John Ashburner,et al.  A fast diffeomorphic image registration algorithm , 2007, NeuroImage.

[48]  S. Hillyard,et al.  Temporal dynamics of selective attention during dichotic listening. , 2009, Cerebral cortex.

[49]  Michael J. Crosse,et al.  Congruent Visual Speech Enhances Cortical Entrainment to Continuous Auditory Speech in Noise-Free Conditions , 2015, The Journal of Neuroscience.

[50]  Jonathan E. Peelle,et al.  Adjusting for global effects in voxel-based morphometry: Gray matter decline in normal aging , 2012, NeuroImage.

[51]  Thomas Lunner,et al.  Single-channel in-ear-EEG detects the focus of auditory attention to concurrent tone streams and mixed speech , 2016, bioRxiv.

[52]  Lawrence M. Ward,et al.  Electrical Neuroimaging of Voluntary Audiospatial Attention: Evidence for a Supramodal Attention Control Network , 2011, The Journal of Neuroscience.

[53]  Thomas Lunner,et al.  Neural tracking of attended versus ignored speech is differentially affected by hearing loss. , 2017, Journal of neurophysiology.

[54]  Barbara G. Shinn-Cunningham,et al.  Auditory Selective Attention Reveals Preparatory Activity in Different Cortical Regions for Selection Based on Source Location and Source Pitch , 2012, Front. Neurosci..

[55]  G. Fink,et al.  Dorsal and Ventral Attention Systems: Distinct Neural Circuits but Collaborative Roles , 2013 .

[56]  Karl J. Friston,et al.  Voxel-based morphometry of the human brain: Methods and applications , 2005 .

[57]  R. Zatorre,et al.  Selective Entrainment of Theta Oscillations in the Dorsal Stream Causally Enhances Auditory Working Memory Performance , 2017, Neuron.