A Graphical Model for Online Auditory Scene Modulation Using EEG Evidence for Attention

Recent findings indicate that brain interfaces have the potential to enable attention-guided auditory scene analysis and manipulation in applications, such as hearing aids and augmented/virtual environments. Specifically, noninvasively acquired electroencephalography (EEG) signals have been demonstrated to carry some evidence regarding, which of multiple synchronous speech waveforms the subject attends to. In this paper, we demonstrate that: 1) using data- and model-driven cross-correlation features yield competitive binary auditory attention classification results with at most 20 s of EEG from 16 channels or even a single well-positioned channel; 2) a model calibrated using equal-energy speech waveforms competing for attention could perform well on estimating attention in closed-loop unbalanced-energy speech waveform situations, where the speech amplitudes are modulated by the estimated attention posterior probability distribution; 3) such a model would perform even better if it is corrected (linearly, in this instance) based on EEG evidence dependence on speech weights in the mixture; and 4) calibrating a model based on population EEG could result in acceptable performance for new individuals/users; therefore, EEG-based auditory attention classifiers may generalize across individuals, leading to reduced or eliminated calibration time and effort.

[1]  E. C. Cmm,et al.  on the Recognition of Speech, with , 2008 .

[2]  Alexander Bertrand,et al.  Auditory-Inspired Speech Envelope Extraction Methods for Improved EEG-Based Auditory Attention Detection in a Cocktail Party Scenario , 2017, IEEE Transactions on Neural Systems and Rehabilitation Engineering.

[3]  D. Abrams,et al.  Right-Hemisphere Auditory Cortex Is Dominant for Coding Syllable Patterns in Speech , 2008, The Journal of Neuroscience.

[4]  N. Mesgarani,et al.  Selective cortical representation of attended speaker in multi-talker speech perception , 2012, Nature.

[5]  T. Picton,et al.  Human Cortical Responses to the Speech Envelope , 2008, Ear and hearing.

[6]  Barbara G. Shinn-Cunningham,et al.  Auditory Selective Attention Reveals Preparatory Activity in Different Cortical Regions for Selection Based on Source Location and Source Pitch , 2012, Front. Neurosci..

[7]  M. D'Zmura,et al.  Envelope responses in single-trial EEG indicate attended speaker in a ‘cocktail party’ , 2014, Journal of Neural Engineering.

[8]  John J. Foxe,et al.  Attentional Selection in a Cocktail Party Environment Can Be Decoded from Single-Trial EEG. , 2015, Cerebral cortex.

[9]  Maarten De Vos,et al.  Decoding the attended speech stream with multi-channel EEG: implications for online, daily-life applications , 2015, Journal of neural engineering.

[10]  J. C. Middlebrooks,et al.  Spatial Stream Segregation by Auditory Cortical Neurons , 2013, The Journal of Neuroscience.

[11]  J. Simon,et al.  Cortical entrainment to continuous speech: functional roles and interpretations , 2014, Front. Hum. Neurosci..

[12]  Ying-Yee Kong,et al.  Differential modulation of auditory responses to attended and unattended speech in different listening conditions , 2014, Hearing Research.

[13]  D. Poeppel,et al.  Phase Patterns of Neuronal Responses Reliably Discriminate Speech in Human Auditory Cortex , 2007, Neuron.

[14]  Anne Sullivan,et al.  Auditory perception. , 1973, British medical journal.

[15]  Christopher K. Kovach,et al.  Temporal Envelope of Time-Compressed Speech Represented in the Human Auditory Cortex , 2009, The Journal of Neuroscience.

[16]  B. Shinn-Cunningham Object-based auditory and visual attention , 2008, Trends in Cognitive Sciences.

[17]  E Ahissar,et al.  Speech comprehension is correlated with temporal response patterns recorded from auditory cortex , 2001, Proceedings of the National Academy of Sciences of the United States of America.

[18]  J. Simon,et al.  Neural coding of continuous speech in auditory cortex during monaural and dichotic listening. , 2012, Journal of neurophysiology.

[19]  Brian N. Pasley,et al.  Reconstructing Speech from Human Auditory Cortex , 2012, PLoS biology.

[20]  B. Shinn-Cunningham,et al.  Selective Attention in Normal and Impaired Hearing , 2008, Trends in amplification.

[21]  Adrian K. C. Lee,et al.  Using neuroimaging to understand the cortical mechanisms of auditory selective attention , 2014, Hearing Research.

[22]  Alessandro Presacco,et al.  Robust decoding of selective auditory attention from MEG in a competing-speaker environment via state-space modeling , 2016, NeuroImage.

[23]  A. Bronkhorst The cocktail-party problem revisited: early processing and selection of multi-talker speech , 2015, Attention, Perception, & Psychophysics.

[24]  Murat Akçakaya,et al.  Toward a brain interface for tracking attended auditory sources , 2016, 2016 IEEE 26th International Workshop on Machine Learning for Signal Processing (MLSP).

[25]  Lee M. Miller,et al.  Auditory attentional control and selection during cocktail party listening. , 2010, Cerebral cortex.

[26]  H. Steven Colburn,et al.  Cortical Transformation of Spatial Processing for Solving the Cocktail Party Problem: A Computational Model123 , 2016, eNeuro.

[27]  Barbara G. Shinn-Cunningham,et al.  Contributions of Sensory Coding and Attentional Control to Individual Differences in Performance in Spatial Auditory Selective Attention Tasks , 2016, Front. Hum. Neurosci..