Development of auditory-evoked reflexes: Visuo-acoustic cues integration in a binocular head

Abstract The goal of this paper is to propose a biologically plausible, functional model of the acquisition of visual, acoustic and multi-modal motor responses. Within this context visual and acoustic spatial cues are considered, fused in a coherent percept and eventually employed to control the orienting behavior of a humanoid robot. The rationale of the approach lies in the possibility to test and empirically prove the correctness of the model through the embodiment and the real interaction of the system with the environment. The model takes into account the fact that (i) acoustic and visual cues are represented with respect to different coordinate frames (head-centric versus retino-centric) and consequently they need to be “aligned” in order to be properly fused, (ii) a teaching signal has to be generated in order to inform the system that the motor performance is not adequate to perform the task (i.e. orient toward the stimulus) and thus adaptation is required, and (iii) vision plays a major role in driving the acquisition of the appropriate map of space but other sources of feedback might be employed as well.

[1]  Gerald M. Edelman,et al.  Adaptation of orienting behavior: from the barn owl to a robotic system , 1999, IEEE Trans. Robotics Autom..

[2]  E. Knudsen The Hearing of the Barn Owl , 1981 .

[3]  Giorgio Metta,et al.  BabyRoBot: A study in sensori-motor development , 1999 .

[4]  Giulio Sandini,et al.  Visuo-inertial stabilization in space-variant binocular systems , 2000, Robotics Auton. Syst..

[5]  E D Young,et al.  Neural network models of sound localization based on directional filtering by the pinna. , 1992, The Journal of the Acoustical Society of America.

[6]  Brian Scassellati,et al.  Humanoid Robots: A New Kind of Tool , 2000, IEEE Intell. Syst..

[7]  Paul M. Hofman,et al.  Relearning sound localization with new ears , 1998, Nature Neuroscience.

[8]  M. Kawato,et al.  A hierarchical neural-network model for control and learning of voluntary movement , 2004, Biological Cybernetics.

[9]  G. Carter,et al.  The generalized correlation method for estimation of time delay , 1976 .

[10]  M Kuperstein,et al.  Neural model of adaptive hand-eye coordination for single postures. , 1988, Science.

[11]  Francisco Vaz,et al.  Perception maps for the local navigation of a mobile robot: a neural network approach , 1994, Proceedings of the 1994 IEEE International Conference on Robotics and Automation.

[12]  E I Knudsen,et al.  Vision-independent adjustment of unit tuning to sound localization cues in response to monaural occlusion in developing owl optic tectum , 1992, The Journal of neuroscience : the official journal of the Society for Neuroscience.

[13]  O. Manolov,et al.  Indoor mobile robot control for environment information gleaning , 2000, Proceedings of the IEEE Intelligent Vehicles Symposium 2000 (Cat. No.00TH8511).

[14]  Patrick Rives,et al.  A new approach to visual servoing in robotics , 1992, IEEE Trans. Robotics Autom..

[15]  M. Cynader,et al.  A computational theory of spectral cue localization , 1993 .

[16]  R. Duda,et al.  Combined monaural and binaural localization of sound sources , 1995, Conference Record of The Twenty-Ninth Asilomar Conference on Signals, Systems and Computers.

[17]  Jie Huang,et al.  A biomimetic system for localization and separation of multiple sound sources , 1994 .

[18]  P M Hofman,et al.  Spectro-temporal factors in two-dimensional human sound localization. , 1998, The Journal of the Acoustical Society of America.

[19]  Michele Rucci,et al.  Theoretical study , 2022 .

[20]  Jie Huang,et al.  A model-based sound localization system and its application to robot navigation , 1999, Robotics Auton. Syst..

[21]  A. Bernardino,et al.  Binocular Visual Tracking : Integration of Perception and Control , 1999 .

[22]  D. M. Green,et al.  Directional sensitivity of sound-pressure levels in the human ear canal. , 1989, The Journal of the Acoustical Society of America.

[23]  B. Stein,et al.  Visual, auditory, and somatosensory convergence on cells in superior colliculus results in multisensory integration. , 1986, Journal of neurophysiology.

[24]  E I Knudsen,et al.  Vision guides the adjustment of auditory localization in young barn owls. , 1985, Science.

[25]  Giulio Sandini,et al.  An anthropomorphic retina-like structure for scene analysis , 1980 .

[26]  Alexandre Bernardino,et al.  Binocular tracking: integrating perception and control , 1999, IEEE Trans. Robotics Autom..

[27]  M Konishi,et al.  Listening with two ears. , 1993, Scientific American.

[28]  Eric I. Knudsen,et al.  A Connectionist Model of the Owl's Sound Localization System , 1993, NIPS.

[29]  C H Keller,et al.  Binaural Cross-Correlation Predicts the Responses of Neurons in the Owl’s Auditory Space Map under Conditions Simulating Summing Localization , 1996, The Journal of Neuroscience.

[30]  Noboru Ohnishi,et al.  Learning Sound Source Localization by Interaction between Motion and Sensing - In a Non-Linear Motor System , 1998, ICONIP.

[31]  E. Knudsen,et al.  Experience-dependent plasticity in the inferior colliculus: a site for visual calibration of the neural representation of auditory space in the barn owl , 1993, The Journal of neuroscience : the official journal of the Society for Neuroscience.

[32]  J. Blauert Spatial Hearing: The Psychophysics of Human Sound Localization , 1983 .

[33]  T. Michael Knasel,et al.  Robotics and autonomous systems , 1988, Robotics Auton. Syst..

[34]  Sjsu ScholarWorks,et al.  Estimating azimuth from speech in a natural auditory environment , 1996 .

[35]  E I Knudsen,et al.  Neural derivation of sound source location: resolution of spatial ambiguities in binaural cues. , 1992, The Journal of the Acoustical Society of America.

[36]  Robert E Irie,et al.  Robust Sound Localization: An Application of an Auditory Perception System for a Humanoid Robot , 1995 .