Real-time Sound Localization With a Binaural Head-system Using a Biologically-inspired Cue-triple Mapping

We present a sound localization system that operates in real-time, calculates three binaural cues (IED, UD, and ITD) and integrates them in a biologically inspired fashion to a combined localization estimation. Position information is furthermore integrated over frequency channels and time. The localization system controls a head motor to fovealize on and track the dominant sound source. Due to an integrated noise-reduction module the system shows robust localization capabilities even in noisy conditions. Real-time performance is gained by multi-threaded parallel operation across different machines using a timestamp-based synchronization scheme to compensate for processing delays

[1]  Brian R Glasberg,et al.  Derivation of auditory filter shapes from notched-noise data , 1990, Hearing Research.

[2]  Volker Willert,et al.  A Probabilistic Model for Binaural Sound Localization , 2006, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[3]  Martin Heckmann,et al.  Auditory Inspired Binaural Robust Sound Source Localization in Echoic and Noisy Environments , 2006, 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[4]  Julian Eggert,et al.  Integrated Research and Development Environment for Real-Time Distributed Embodied Intelligent Systems , 2006, 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[5]  Malcolm Slaney,et al.  An Efficient Implementation of the Patterson-Holdsworth Auditory Filter Bank , 1997 .

[6]  Hiroshi Mizoguchi,et al.  Detecting and segmenting sound sources by using microphone array , 2004, 4th IEEE/RAS International Conference on Humanoid Robots, 2004..

[7]  Erik Berglund,et al.  Sound source localisation through active audition , 2005, 2005 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[8]  S. Boll,et al.  Suppression of acoustic noise in speech using spectral subtraction , 1979 .

[9]  Martin Heckmann,et al.  Sound source separation for a robot based on pitch , 2005, 2005 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[10]  DeLiang Wang,et al.  Speech segregation based on sound localization , 2001, IJCNN'01. International Joint Conference on Neural Networks. Proceedings (Cat. No.01CH37222).

[11]  Hiroaki Kitano,et al.  Sound and Visual Tracking for Humanoid Robot , 2004, Applied Intelligence.

[12]  Rhee Man Kil,et al.  Sound segregation based on binaural zero-crossings , 2005, Interspeech.

[13]  Jean Rouat,et al.  Making a robot recognize three simultaneous sentences in real-time , 2005, 2005 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[14]  Kazuhiro Nakadai,et al.  Sound source tracking with directivity pattern estimation using a 64 ch microphone array , 2005, 2005 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[15]  Hideharu Amano,et al.  Implementation of active direction-pass filter on dynamically reconfigurable processor , 2005, 2005 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[16]  Israel Cohen,et al.  Noise spectrum estimation in adverse environments: improved minima controlled recursive averaging , 2003, IEEE Trans. Speech Audio Process..

[17]  Heiko Wersing,et al.  Peripersonal space and object recognition for humanoids , 2005, 5th IEEE-RAS International Conference on Humanoid Robots, 2005..

[18]  I. Cohen,et al.  Noise estimation by minima controlled recursive averaging for robust speech enhancement , 2002, IEEE Signal Processing Letters.