A Joint Auditory Attention Decoding and Adaptive Binaural Beamforming Algorithm for Hearing Devices

Traditional adaptive binaural beamforming algorithms for hearing devices often assume that the target talker is known or can be derived from the listener’s look direction. When this assumption is violated, the traditional beamforming algorithms often produce distorted target speech and less than optimal noise and interference suppression. Recent advances in electroencephalography (EEG) and its applications to auditory attention decoding have offered a potential solution for tracking the listeners auditory attention in a multi-talker environment [2]–[5]. In this paper, we propose a unified model for joint auditory attention decoding and adaptive binaural beamforming, and solve the problem using an iterative optimization approach. The proposed algorithm has two advantages over the existing algorithms. First, the optimization objective aims to balance auditory attention alignment, target speech distortion, noise and interference suppression. Secondly, there is no need to estimate the speech envelope of each talker from the noisy and reverberant mixture which is a very challenging problem in practice. The proposed algorithm was evaluated using a newly recorded EEG database for a multi-talker, noisy and reverberant environment [6]. The evaluation results confirm the benefits of the proposed algorithm.

[1]  Marc Moonen,et al.  Acoustic Beamforming for Hearing Aid Applications , 2010 .

[2]  Marc Moonen,et al.  Robustness analysis of multichannel Wiener filtering and generalized sidelobe cancellation for multimicrophone noise reduction in hearing aid applications , 2005, IEEE Transactions on Speech and Audio Processing.

[3]  Brian N. Pasley,et al.  Reconstructing Speech from Human Auditory Cortex , 2012, PLoS biology.

[4]  John J. Foxe,et al.  At what time is the cocktail party? A late locus of selective attention to natural speech , 2012, The European journal of neuroscience.

[5]  Tao Zhang,et al.  An optimization model for electroencephalography-assisted binaural beamforming , 2018 .

[6]  Zhuo Chen,et al.  Neural decoding of attentional selection in multi-speaker environments without access to clean sources , 2017, Journal of neural engineering.

[7]  M. Elhilali,et al.  The Auditory System at the Cocktail Party , 2017 .

[8]  Alexander Bertrand,et al.  EEG-based auditory attention detection: boundary conditions for background noise and speaker positions. , 2018, Journal of neural engineering.

[9]  Zhi-Quan Luo,et al.  On the linear convergence of the alternating direction method of multipliers , 2012, Mathematical Programming.

[10]  Ruth Y Litovsky,et al.  The benefit of binaural hearing in a cocktail party: effect of location and type of interferer. , 2004, The Journal of the Acoustical Society of America.

[11]  Aaas News,et al.  Book Reviews , 1893, Buffalo Medical and Surgical Journal.

[12]  Dimitri P. Bertsekas,et al.  Nonlinear Programming , 1997 .

[13]  John J. Foxe,et al.  Attentional Selection in a Cocktail Party Environment Can Be Decoded from Single-Trial EEG. , 2015, Cerebral cortex.

[14]  S. Marple Computing the discrete-time 'analytic' signal via FFT , 1997 .

[15]  Jonathan Z. Simon,et al.  Real-Time Tracking of Selective Auditory Attention From M/EEG: A Bayesian Filtering Approach , 2017, bioRxiv.

[16]  Alexander Bertrand,et al.  Auditory-Inspired Speech Envelope Extraction Methods for Improved EEG-Based Auditory Attention Detection in a Cocktail Party Scenario , 2017, IEEE Transactions on Neural Systems and Rehabilitation Engineering.

[17]  Torsten Dau,et al.  Noise-robust cortical tracking of attended speech in real-world acoustic scenes , 2017, NeuroImage.

[18]  Alain de Cheveigné,et al.  Decoding the auditory brain with canonical component analysis , 2017, NeuroImage.

[19]  Simon Haykin,et al.  The Cocktail Party Problem , 2005, Neural Computation.

[20]  Sharon Gannot,et al.  The Binaural LCMV Beamformer and its Performance Analysis , 2016, IEEE/ACM Transactions on Audio, Speech, and Language Processing.

[21]  Sven Nordholm,et al.  Multichannel Signal Enhancement Algorithms for Assisted Listening Devices: Exploiting spatial diversity using multiple microphones , 2015, IEEE Signal Processing Magazine.

[22]  Tao Zhang,et al.  A penalized inequality-constrained minimum variance beamformer with applications in hearing aids , 2017, 2017 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA).

[23]  W. Marsden I and J , 2012 .

[24]  Maarten De Vos,et al.  Decoding the attended speech stream with multi-channel EEG: implications for online, daily-life applications , 2015, Journal of neural engineering.