Beamspace blind signal separation for speech enhancement

Signal processing methods for speech enhancement are of vital interest for communications equipments. In particular, multichannel algorithms, which perform spatial filtering to separate signals that have overlapping frequency content but different spatial origins, are important for a wide range of applications. Two of the most popular multichannel methods are blind signal separation (BSS) and beamforming. Briefly, (BSS) separates mixed sources by optimizing the statistical independence among the outputs whilst beamforming optimizes the look direction of the desired source(s). However, both methods have separation limitations, in that BSS succumbs to reverberant environments and beamforming is very sensitive to array model mismatch. In this paper, we propose a novel hybrid scheme, called beamspace BSS, which is intended to compensate the aforementioned separation weaknesses by jointly optimizing the spatial selectivity and statistical independence of the sources. We show that beamspace BSS outperforms the separation performance of the conventional sensor space BSS significantly, particularly in reverberant room environments.

[1]  Michael S. Brandstein,et al.  Microphone Arrays - Signal Processing Techniques and Applications , 2001, Microphone Arrays.

[2]  A. Gershman,et al.  Beamspace preprocessing with an improved robustness against out-of-sector sources using second-order cone programming , 2004, Processing Workshop Proceedings, 2004 Sensor Array and Multichannel Signal.

[3]  Meir Feder,et al.  Multi-channel signal separation by decorrelation , 1993, IEEE Trans. Speech Audio Process..

[4]  Andrzej Cichocki,et al.  Neural networks for blind decorrelation of signals , 1997, IEEE Trans. Signal Process..

[5]  P. Peterson Simulating the response of multiple microphones to a single acoustic source in a reverberant room. , 1986, The Journal of the Acoustical Society of America.

[6]  Shoko Araki,et al.  Time domain blind source separation of non-stationary convolved signals by utilizing geometric beamforming , 2002, Proceedings of the 12th IEEE Workshop on Neural Networks for Signal Processing.

[7]  Andrzej Cichocki,et al.  Adaptive blind signal and image processing , 2002 .

[8]  Kevin Buckley,et al.  Spatial-spectrum estimation in a location sector , 1990, IEEE Trans. Acoust. Speech Signal Process..

[9]  Charles R. Johnson,et al.  Matrix analysis , 1985, Statistical Inference for Engineers and Data Scientists.

[10]  Zohra Yermeche,et al.  SOFT-CONSTRAINED SUBBAND BEAMFORMING FOR SPEECH ENHANCEMENT , 2007 .

[11]  Lucas C. Parra,et al.  Convolutive blind separation of non-stationary sources , 2000, IEEE Trans. Speech Audio Process..

[12]  Barry D. Van Veen,et al.  MEG and EEG source localization in beamspace , 2006, IEEE Transactions on Biomedical Engineering.

[13]  B. M. Eyuboglu,et al.  Incorporating a-priori information in EIT , 1995, Proceedings of 17th International Conference of the Engineering in Medicine and Biology Society.

[14]  H.L. Van Trees,et al.  Beamspace MODE , 2001, Conference Record of Thirty-Fifth Asilomar Conference on Signals, Systems and Computers (Cat.No.01CH37256).

[15]  Harry B. Lee,et al.  Resolution threshold of beamspace MUSIC for two closely spaced emitters , 1990, IEEE Trans. Acoust. Speech Signal Process..

[16]  Sven Nordholm,et al.  Adaptive microphone array employing calibration signals: an analytical evaluation , 1999, IEEE Trans. Speech Audio Process..

[17]  Dennis R. Morgan,et al.  A beamforming approach to permutation alignment for multichannel frequency-domain blind speech separation , 2002, 2002 IEEE International Conference on Acoustics, Speech, and Signal Processing.

[18]  Christopher V. Alvino,et al.  Geometric source separation: merging convolutive source separation with geometric beamforming , 2001, Neural Networks for Signal Processing XI: Proceedings of the 2001 IEEE Signal Processing Society Workshop (IEEE Cat. No.01TH8584).

[19]  Lucas C. Parra,et al.  The generalized sidelobe decorrelator , 2001, Proceedings of the 2001 IEEE Workshop on the Applications of Signal Processing to Audio and Acoustics (Cat. No.01TH8575).

[20]  L. J. Griffiths,et al.  An alternative approach to linearly constrained adaptive beamforming , 1982 .

[21]  Moeness G. Amin,et al.  Minimum bias spatial filters for beamspace direction-of-arrival estimation , 1998 .

[22]  E. C. Cmm,et al.  on the Recognition of Speech, with , 2008 .

[23]  Shoko Araki,et al.  The fundamental limitation of frequency domain blind source separation for convolutive mixtures of speech , 2003, IEEE Trans. Speech Audio Process..

[24]  Hiroshi Sawada,et al.  A robust and precise method for solving the permutation problem of frequency-domain blind source separation , 2004, IEEE Transactions on Speech and Audio Processing.

[25]  Shoji Makino,et al.  Blind Source Separation of Convolutive Mixtures of Speech , 2003 .

[26]  Jacob Benesty,et al.  Speech Enhancement , 2010 .

[27]  Simon J. Godsill,et al.  Statistical reconstruction and analysis of autoregressive signals in impulsive noise using the Gibbs sampler , 1998, IEEE Trans. Speech Audio Process..

[28]  Petre Stoica,et al.  Incorporating a priori information into MUSIC-algorithms and analysis , 1995, Signal Process..

[29]  Jean-Francois Cardoso,et al.  Blind signal separation: statistical principles , 1998, Proc. IEEE.