Autonomous and Interactive Improvement of Binocular Visual Depth Estimation through Sensorimotor Interaction

We investigate how a humanoid robot with a randomly initialized binocular vision system can learn to improve judgments about egocentric distances using limited action and interaction that might be available to human infants. First, we show how distance estimation can be improved autonomously. We consider our approach to be autonomous because the robot learns to accurately estimate distance without a human teacher providing the distances to training targets. We find that actions that, in principle, do not alter the robot's distance to the target are a powerful tool for exposing estimation errors. These errors can be used to train a distance estimator. Furthermore, the simple action used (i.e., neck rotation) does not require high level cognitive processing or fine motor skill. Next, we investigate how interaction with humans can further improve visual distance estimates. We find that human interaction can improve distance estimates for far targets outside of the robot's peripersonal space. This is accomplished by extending our autonomous approach above to integrate additional information provided by a human. Together these experiments suggest that both action and interaction are important tools for improving perceptual estimates.

[1]  R Held,et al.  The development of eye alignment, convergence, and sensory binocularity in young infants. , 1994, Investigative ophthalmology & visual science.

[2]  B. Rogers,et al.  Similarities between motion parallax and stereopsis in human depth perception , 1982, Vision Research.

[3]  Christof Koch,et al.  A Model of Saliency-Based Visual Attention for Rapid Scene Analysis , 2009 .

[4]  Minho Lee,et al.  Biologically motivated vergence control system using human-like selective attention model , 2006, Neurocomputing.

[5]  Jürgen Schmidhuber,et al.  Formal Theory of Creativity, Fun, and Intrinsic Motivation (1990–2010) , 2010, IEEE Transactions on Autonomous Mental Development.

[6]  G. Rizzolatti,et al.  Action recognition in the premotor cortex. , 1996, Brain : a journal of neurology.

[7]  KochChristof,et al.  A Model of Saliency-Based Visual Attention for Rapid Scene Analysis , 1998 .

[8]  Barbara Caputo,et al.  Using Object Affordances to Improve Object Recognition , 2011, IEEE Transactions on Autonomous Mental Development.

[9]  Stephen P. Boyd,et al.  Convex Optimization , 2004, Algorithms and Theory of Computation Handbook.

[10]  S. Chieffi,et al.  Visual illusion and action , 1996, Neuropsychologia.

[11]  Mark Wexler,et al.  Depth perception by the active observer , 2005, Trends in Cognitive Sciences.

[12]  Karl Kral,et al.  Side-to-side head movements to obtain motion depth cues: A short review of research on the praying mantis , 1998, Behavioural Processes.

[13]  Juyang Weng,et al.  Brain-Like Emergent Spatial Processing , 2012, IEEE Transactions on Autonomous Mental Development.

[14]  N. Kanwisher,et al.  fMRI Adaptation Reveals Mirror Neurons in Human Inferior Parietal Cortex , 2008, Current Biology.

[15]  Richard L. Lewis,et al.  Intrinsically Motivated Reinforcement Learning: An Evolutionary Perspective , 2010, IEEE Transactions on Autonomous Mental Development.

[16]  Benjamin Kuipers,et al.  Bootstrap learning of foundational representations , 2006, Connect. Sci..

[17]  R. Held,et al.  MOVEMENT-PRODUCED STIMULATION IN THE DEVELOPMENT OF VISUALLY GUIDED BEHAVIOR. , 1963, Journal of comparative and physiological psychology.

[18]  Paul R. Schrater,et al.  How Haptic Size Sensations Improve Distance Perception , 2011, PLoS Comput. Biol..

[19]  A. Treisman,et al.  A feature-integration theory of attention , 1980, Cognitive Psychology.

[20]  Mark Nawrot,et al.  The development of depth perception from motion parallax in infancy , 2009, Attention, perception & psychophysics.

[21]  Minho Lee,et al.  Stereo saliency map considering affective factors and selective motion analysis in a dynamic environment , 2008, Neural Networks.

[22]  M. Goodale,et al.  Size-contrast illusions deceive the eye but not the hand , 1995, Current Biology.

[23]  R. Passingham,et al.  Action observation and acquired motor skills: an FMRI study with expert dancers. , 2005, Cerebral cortex.

[24]  Karl Kral,et al.  Behavioural–analytical studies of the role of head movements in depth perception in insects, birds and mammals , 2003, Behavioural Processes.

[25]  B JULESZ,et al.  Binocular Depth Perception without Familiarity Cues , 1964, Science.

[26]  A. S. Gilinsky Perceived size and distance in visual space. , 1951, Psychological review.

[27]  T. Poggio,et al.  The analysis of stereopsis. , 1984, Annual review of neuroscience.

[28]  Michael Brady,et al.  Saliency, Scale and Image Description , 2001, International Journal of Computer Vision.

[29]  D. Carey,et al.  More thoughts on perceiving and grasping the Müller–Lyer illusion , 1999, Neuropsychologia.

[30]  Darrell Whitley,et al.  A genetic algorithm tutorial , 1994, Statistics and Computing.

[31]  Daniel Chern-Yeow Eng,et al.  Autonomous Learning of the Semantics of Internal Sensory States Based on Motor Exploration , 2007, Int. J. Humanoid Robotics.

[32]  Benjamin Kuipers,et al.  Autonomous Learning of High-Level States and Actions in Continuous Environments , 2012, IEEE Transactions on Autonomous Mental Development.

[33]  Minho Lee,et al.  Saliency map model with adaptive masking based on independent component analysis , 2002, Neurocomputing.