Theoretical framework for the design of microphone arrays for robot audition

An important part of a human-like robot is robot audition. Previous work presented systems capable of sound localization and source segregation based on microphone arrays of various configurations. However, no theoretical framework for assessing the quality of these array configurations has been presented. In the current paper such a measure is proposed based on the generalized HRTFs that account for microphone positions other than the ears. The measure is analyzed theoretically with respect to beamforming robustness and DOA estimation accuracy. The measure is then used to find the optimal location of a single microphone and a pair of microphones based on the generalized HRTF database obtained by means of BEM simulation. The results are not surprising, showing that the best position of a single microphone is the ear canal. For a pair of microphones, the results generally show that the sensors should be maximally spatially separated.

[1]  Brian F. G. Katz,et al.  Round Robin Comparison of HRTF Simulation Systems: Preliminary Results , 2007 .

[2]  Pawel Strumillo,et al.  Advances in Sound Localization , 2011 .

[3]  Simon N. Chandler-Wilde,et al.  Boundary element methods for acoustics , 2007 .

[4]  Martin Heckmann,et al.  Teaching a humanoid robot: Headset-free speech interaction for audio-visual association learning , 2009, RO-MAN 2009 - The 18th IEEE International Symposium on Robot and Human Interactive Communication.

[5]  Petre Stoica,et al.  MUSIC, maximum likelihood, and Cramer-Rao bound , 1989, IEEE Transactions on Acoustics, Speech, and Signal Processing.

[6]  B F Katz,et al.  Acoustic absorption measurement of human hair and skin within the audible frequency range. , 2000, The Journal of the Acoustical Society of America.

[7]  D. M. Green,et al.  Sound localization by human listeners. , 1991, Annual review of psychology.

[8]  N. J. A. Sloane,et al.  McLaren’s improved snub cube and other new spherical designs in three dimensions , 1996, Discret. Comput. Geom..

[9]  José Santos-Victor,et al.  Sound Localization for Humanoid Robots - Building Audio-Motor Maps based on the HRTF , 2006, 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[10]  B. Rafaely,et al.  Objective measure for sound localization based on head-related transfer functions , 2012, 2012 IEEE 27th Convention of Electrical and Electronics Engineers in Israel.

[11]  Harry L. Van Trees,et al.  Optimum Array Processing: Part IV of Detection, Estimation, and Modulation Theory , 2002 .

[12]  Klaus Diepold,et al.  A Novel Humanoid Binaural 3D Sound Localization and Separation Algorithm , 2006, 2006 6th IEEE-RAS International Conference on Humanoid Robots.

[13]  Lawrence D. Rosenblum,et al.  Hearing space: Identifying rooms by reflected sound , 2005 .

[14]  Boaz Rafaely,et al.  Optimal Real-Weighted Beamforming With Application to Linear and Spherical Arrays , 2012, IEEE Transactions on Audio, Speech, and Language Processing.

[15]  Karsten Berns,et al.  3D Audio Perception System for Humanoid Robots , 2009, 2009 Second International Conferences on Advances in Computer-Human Interactions.

[16]  Md Akhtaruzzaman,et al.  Evolution of Humanoid Robot and contribution of various countries in advancing the research and development of the platform , 2010, ICCAS 2010.

[17]  Karim Abed-Meraim,et al.  Adaptive blind source separation with HRTFs beamforming preprocessing , 2012, 2012 IEEE 7th Sensor Array and Multichannel Signal Processing Workshop (SAM).

[18]  R. O. Schmidt,et al.  Multiple emitter location and signal Parameter estimation , 1986 .

[19]  Harry L. Van Trees,et al.  Optimum Array Processing , 2002 .

[20]  Ruth Y. Litovsky,et al.  Spatial Release from Masking , 2012 .