Acoustic localization by interaural level difference

Interaural level difference (ILD) is an important cue for acoustic localization. Although its behavior has been studied extensively in natural systems, it remains an untapped resource for computer-based systems. We investigate the possibility of using ILD for acoustic localization, deriving constraints on the location of a sound source given the relative energy level of the signals received by two microphones. We then present an algorithm for computing the sound source location by combining likelihood functions, one for each microphone pair. Experimental results show that accurate acoustic localization can be achieved using ILD alone.

[1]  M S Brandstein Time-delay estimation of reverberated speech exploiting harmonic structure. , 1999, The Journal of the Acoustical Society of America.

[2]  Stanley T. Birchfield A unifying framework for acoustic localization , 2004, 2004 12th European Signal Processing Conference.

[3]  Maurizio Omologo,et al.  Acoustic source location in a three-dimensional space using crosspower spectrum phase , 1997, 1997 IEEE International Conference on Acoustics, Speech, and Signal Processing.

[4]  Andrew Blake,et al.  Nonlinear filtering for speaker tracking in noisy and reverberant environments , 2001, 2001 IEEE International Conference on Acoustics, Speech, and Signal Processing. Proceedings (Cat. No.01CH37221).

[5]  Darren B. Ward,et al.  Particle filter beamforming for acoustic source localization in a reverberant environment , 2002, 2002 IEEE International Conference on Acoustics, Speech, and Signal Processing.

[6]  Benoît Champagne,et al.  Cepstral prefiltering for time delay estimation in reverberant environments , 1995, 1995 International Conference on Acoustics, Speech, and Signal Processing.

[7]  Maurizio Omologo,et al.  Use of the crosspower-spectrum phase in acoustic event location , 1997, IEEE Trans. Speech Audio Process..

[8]  J. Flanagan,et al.  Computer‐steered microphone arrays for sound transduction in large rooms , 1985 .

[9]  Jont B. Allen,et al.  Image method for efficiently simulating small‐room acoustics , 1976 .

[10]  Stanley T. Birchfield,et al.  Acoustic source direction by hemisphere sampling , 2001, 2001 IEEE International Conference on Acoustics, Speech, and Signal Processing. Proceedings (Cat. No.01CH37221).

[11]  Stanley T. Birchfield,et al.  Fast Bayesian acoustic localization , 2002, 2002 IEEE International Conference on Acoustics, Speech, and Signal Processing.

[12]  D.N.L. Howell Spatial hearing: The psychophysics of human sound localization 1983: by Jens Blauert (translated by John S. Allen). Cambridge, Massachusetts: MIT Press. 408 pp. Price £36.00. ISBN 0 262 02190 0 , 1985 .

[13]  W. Yost Fundamentals of hearing: An introduction , 1977 .

[14]  Michael S. Brandstein,et al.  A practical methodology for speech source localization with microphone arrays , 1997, Comput. Speech Lang..

[15]  Larry S. Davis,et al.  Active speech source localization by a dual coarse-to-fine search , 2001, 2001 IEEE International Conference on Acoustics, Speech, and Signal Processing. Proceedings (Cat. No.01CH37221).

[16]  G. Carter,et al.  The generalized correlation method for estimation of time delay , 1976 .

[17]  E I Knudsen,et al.  Neural derivation of sound source location: resolution of spatial ambiguities in binaural cues. , 1992, The Journal of the Acoustical Society of America.