Cross-modal body representation based on visual attention by saliency

In performing various kinds of tasks, body representation is one of the most fundamental issues for physical agents (humans, primates, and robots). Especially during tool-use by Japanese macaque monkeys, neurophysiological evidence shows that the representation can be dynamically reconstructed by spatio-temporal integration of different sensor modalities so that it can be adaptive to environmental changes. However, to construct such a representation, an issue to be solved is how to associate which information among various sensory data. This paper presents a method that constructs cross-modal body representation from vision, touch, and proprioception. When the robot touches something, the activation of tactile sense triggers the construction process of the visual receptive field for body parts that can be found by visual attention based on saliency map and consequently regarded as the end effector. Simultaneously, proprioceptive information is associated with this visual receptive field to construct the cross-modal body representation. The computer simulation results are comparable to the activities of parietal neurons found in the Japanese macaque monkeys. Various conditions are also investigated so that what kind of information is important to generate the same results as findings in neurophysiology.

[1]  吉川 雄一郎 Subjective Robot Imitation by Finding Invariance , 2005 .

[2]  Minoru Asada,et al.  Cooperative Behavior Acquisition for Mobile Robots in Dynamically Changing Real Worlds Via Vision-Based Reinforcement Learning and Development , 1999, Artif. Intell..

[3]  Yasuo Kuniyoshi,et al.  Adaptive body schema for robotic tool-use , 2006, Adv. Robotics.

[4]  A. Maravita,et al.  Tools for the body (schema) , 2004, Trends in Cognitive Sciences.

[5]  E. Knudsen Fundamental components of attention. , 2007, Annual review of neuroscience.

[6]  Laurent Itti,et al.  Realistic avatar eye and head animation using a neurobiological model of visual attention , 2004, SPIE Optics + Photonics.

[7]  Rj Dolan,et al.  Phantoms in the brain: Probing the mysteries of the human mind , 1998 .

[8]  S. Fuke,et al.  VIP neuron model: Head-centered cross-modal representation of the peri-personal space around the face , 2008, 2008 7th IEEE International Conference on Development and Learning.

[9]  M. Tanaka,et al.  Coding of modified body schema during tool use by macaque postcentral neurones. , 1996, Neuroreport.

[10]  Helena De Preester,et al.  Body image and body schema , 2005 .

[11]  Minoru Asada,et al.  Cognitive developmental robotics as a new paradigm for the design of humanoid robots , 2001, Robotics Auton. Syst..

[12]  M. V. Velzen,et al.  Self-organizing maps , 2007 .

[13]  P. Downing,et al.  The neural basis of visual body perception , 2007, Nature Reviews Neuroscience.

[14]  Alexander Stoytchev Toward Video-Guided Robot Behaviors , 2007 .

[15]  Y. Iwamura Hierarchical somatosensory processing , 1998, Current Opinion in Neurobiology.

[16]  G. Holmes,et al.  Sensory disturbances from cerebral lesions , 1911 .