LOCALIZATION BASED SEPARATION OF MOVING AUDIO SOURCES FROM STEREO MIXTURES BY ADAPTIVE BEAMFORMING

The separation of moving audio sources from their mixtures is practically more useful as well as challenging in the area of source separation. In this paper, a microphone pair is used to capture the signals of multiple moving audio sources. It is assumed that the sources are static within a small time slot. Then localization based separation from the mixtures is obtained by applying adaptive beamforming within that slot. A coherence based source localization method is proposed here. When the sources are moving randomly, it is obvious to arise a permutation problem among the sources separated from the successive time slots. A data- adaptive audio source discrimination method is introduced to resolve such problem. Then the overall separation is obtained by concatenating the sources segregated from consecutive time slots. The experimental results are presented to demonstrate the performance of the proposed method.

[1]  G. R. Arce,et al.  Nearfield spot-beamforming with distributed arrays , 2000, Proceedings of the 2000 IEEE Sensor Array and Multichannel Signal Processing Workshop. SAM 2000 (Cat. No.00EX410).

[2]  C. Faller,et al.  Source localization in complex listening situations: selection of binaural cues based on interaural coherence. , 2004, The Journal of the Acoustical Society of America.

[3]  Christopher V. Alvino,et al.  Geometric source separation: merging convolutive source separation with geometric beamforming , 2001, Neural Networks for Signal Processing XI: Proceedings of the 2001 IEEE Signal Processing Society Workshop (IEEE Cat. No.01TH8584).

[4]  Keikichi Hirose,et al.  Multi-band approach of audio source discrimination with empirical mode decomposition , 2005, INTERSPEECH.

[5]  Özgür Yilmaz,et al.  On the approximate W-disjoint orthogonality of speech , 2002, 2002 IEEE International Conference on Acoustics, Speech, and Signal Processing.

[6]  B.D. Van Veen,et al.  Beamforming: a versatile approach to spatial filtering , 1988, IEEE ASSP Magazine.

[7]  Hiroaki Kitano,et al.  Separating three simultaneous speeches with two microphones by integrating auditory and visual processing , 2001, INTERSPEECH.

[8]  DeLiang Wang,et al.  Speech segregation based on sound localization , 2001, IJCNN'01. International Joint Conference on Neural Networks. Proceedings (Cat. No.01CH37222).

[9]  Lars Kai Hansen,et al.  Semi-blind source separation using head-related transfer functions [speech signal separation] , 2004, 2004 IEEE International Conference on Acoustics, Speech, and Signal Processing.