On Foveated Gaze Control and Combined Gaze and Locomotion Planning

This chapter presents recent research results of our laboratory in the area of vision and locomotion coordination with an emphasis on foveated multi-camera vision. A novel active vision planning concept is presented which coordinates the individual devices of a foveated multi-camera system. Gaze direction control is combined with trajectory planning based on information theoretic criteria to provide vision-based autonomous exploring robots with accurate models of their environment. With the help of velocity and yaw angle sensors, mobile robots can update the internal knowledge about their current position and orientation from a previous time step; this process is commonly referred to as dead-reckoning. Due to measurement errors and slippage these estimations are erroneous and position accuracy degrades over time causing a drift of the estimated robot pose. To overcome the drift problem it is common to take absolute measurements evaluating visual information, which are fused dynamically with the odometry data by applying Kalman-filter or other techniques, e.g. (Dissanayake et al., 2001). The use of active vision systems for navigation is state-of-the-art providing a situation-related selective allocation of vision sensor resources, e.g. (Davison & Murray, 2002; Seara et al., 2003; Vidal-Calleja et al., 2006). Active vision systems comprising only one type of vision sensor face a trade-off between field of view and measurement accuracy due to limitations of sensor size and resolution, and of computational resources. In order to overcome this drawback the combined use of several vision devices with different fields of view and measurement accuracies is known which is called foveated, multi-resolution, or multi-focal vision, e.g. cf. (Dickmanns, 2003; Kuhnlenz et al., 2006; Ude et al., 2006). Thereby, the individual vision devices can be independently controlled according to the current situation and task requirements. The use of foveated active vision for humanoid robot navigation is considered novel. Active vision is also frequently utilized in the context of robotic exploration. Yet, gaze control and locomotion planning are generally decoupled in state-of-the-art approaches to simultaneous localization and mapping (SLAM). An integrated locomotion planning and gaze direction control concept maximizing the collected amount of information is presented in the second part of this chapter. This strategy results in more accurate autonomously acquired environment representations and robot position estimates compared to state-of-the-art approaches. The chapter is organized as follows: In Section 2 vision-based localization and mapping in the context of humanoid robots is surveyed; Section 3 is concerned with foveated multi-

[1]  Radu Horaud,et al.  Camera cooperation for achieving visual attention , 2005, Machine Vision and Applications.

[2]  John J. Leonard,et al.  Adaptive Mobile Robot Navigation and Mapping , 1999, Int. J. Robotics Res..

[3]  Sebastian Thrun,et al.  FastSLAM: a factored solution to the simultaneous localization and mapping problem , 2002, AAAI/IAAI.

[4]  Alexei Makarenko,et al.  Information based adaptive robotic exploration , 2002, IEEE/RSJ International Conference on Intelligent Robots and Systems.

[5]  Martin Buss,et al.  Information-Based Gaze Direction Planning Algorithm for SLAM , 2006, 2006 6th IEEE-RAS International Conference on Humanoid Robots.

[6]  Brian Yamauchi,et al.  Frontier-based exploration using multiple robots , 1998, AGENTS '98.

[7]  Andrew J. Davison,et al.  Real-time simultaneous localisation and mapping with a single camera , 2003, Proceedings Ninth IEEE International Conference on Computer Vision.

[8]  C. E. SHANNON,et al.  A mathematical theory of communication , 1948, MOCO.

[9]  James J. Little,et al.  Vision-based global localization and mapping for mobile robots , 2005, IEEE Transactions on Robotics.

[10]  Olivier Stasse,et al.  Real-time 3D SLAM for Humanoid Robot considering Pattern Generator Information , 2006, 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[11]  Howie Choset,et al.  Autonomous exploration via regions of interest , 2003, Proceedings 2003 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2003) (Cat. No.03CH37453).

[12]  Brian Scassellati,et al.  Eye Finding via Face Detection for a Foveated Active Vision System , 1998, AAAI/IAAI.

[13]  Hirochika Inoue,et al.  Using visual odometry to create 3D maps for online footstep planning , 2005, 2005 IEEE International Conference on Systems, Man and Cybernetics.

[14]  James J. Little,et al.  Autonomous vision-based exploration and mapping using hybrid maps and Rao-Blackwellised particle filters , 2006, 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[15]  Trevor Darrell Reinforcement Learning of Active Recognition Behaviors , 1997, NIPS 1997.

[16]  Randall Smith,et al.  Estimating uncertain spatial relationships in robotics , 1986, Proceedings. 1987 IEEE International Conference on Robotics and Automation.

[17]  E. D. Dickmanns,et al.  EMS-vision: gaze control in autonomous vehicles , 2000, Proceedings of the IEEE Intelligent Vehicles Symposium 2000 (Cat. No.00TH8511).

[18]  G LoweDavid,et al.  Distinctive Image Features from Scale-Invariant Keypoints , 2004 .

[19]  Sang Joon Kim,et al.  A Mathematical Theory of Communication , 2006 .

[20]  David W. Murray,et al.  Simultaneous Localization and Map-Building Using Active Vision , 2002, IEEE Trans. Pattern Anal. Mach. Intell..

[21]  Alexander Zelinsky,et al.  Vision In and Out of Vehicles: Integrated Driver and Road Scene Monitoring , 2002, ISER.

[22]  Salah Sukkarieh,et al.  An information-theoretic approach to autonomous navigation and guidance of an uninhabited aerial vehicle in unknown environments , 2005, 2005 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[23]  William Whittaker,et al.  Autonomous exploration using multiple sources of information , 2001, Proceedings 2001 ICRA. IEEE International Conference on Robotics and Automation (Cat. No.01CH37164).

[24]  Evangelos E. Milios,et al.  Globally Consistent Range Scan Alignment for Environment Mapping , 1997, Auton. Robots.

[25]  David G. Lowe,et al.  Shape indexing using approximate nearest-neighbour search in high-dimensional spaces , 1997, Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[26]  Gordon Cheng,et al.  Foveated vision systems with two cameras per eye , 2006, Proceedings 2006 IEEE International Conference on Robotics and Automation, 2006. ICRA 2006..

[27]  W MurrayDavid,et al.  Simultaneous Localization and Map-Building Using Active Vision , 2002 .

[28]  Hanumant Singh,et al.  Visually augmented navigation in an unstructured environment using a delayed state history , 2004, IEEE International Conference on Robotics and Automation, 2004. Proceedings. ICRA '04. 2004.

[29]  Kurt Konolige,et al.  Incremental mapping of large cyclic environments , 1999, Proceedings 1999 IEEE International Symposium on Computational Intelligence in Robotics and Automation. CIRA'99 (Cat. No.99EX375).

[30]  Wolfram Burgard,et al.  Information Gain-based Exploration Using Rao-Blackwellized Particle Filters , 2005, Robotics: Science and Systems.

[31]  Salah Sukkarieh,et al.  Airborne simultaneous localisation and map building , 2003, 2003 IEEE International Conference on Robotics and Automation (Cat. No.03CH37422).

[32]  Martin Buss,et al.  A multi-focal high-performance vision system , 2006, Proceedings 2006 IEEE International Conference on Robotics and Automation, 2006. ICRA 2006..

[33]  Cyrill Stachniss Information Gain-based Exploration , 2009 .

[34]  Hugh F. Durrant-Whyte,et al.  A solution to the simultaneous localization and map building (SLAM) problem , 2001, IEEE Trans. Robotics Autom..

[35]  Teresa A. Vidal-Calleja,et al.  Active control for single camera SLAM , 2006, Proceedings 2006 IEEE International Conference on Robotics and Automation, 2006. ICRA 2006..

[36]  Nando de Freitas,et al.  Rao-Blackwellised Particle Filtering for Dynamic Bayesian Networks , 2000, UAI.

[37]  James J. Little,et al.  Mobile Robot Localization and Mapping with Uncertainty using Scale-Invariant Visual Landmarks , 2002, Int. J. Robotics Res..

[38]  Sven Koenig,et al.  Improved analysis of greedy mapping , 2003, Proceedings 2003 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2003) (Cat. No.03CH37453).

[39]  Peter C. Cheeseman,et al.  Estimating uncertain spatial relationships in robotics , 1986, Proceedings. 1987 IEEE International Conference on Robotics and Automation.

[40]  John J. Leonard,et al.  A Computationally Efficient Method for Large-Scale Concurrent Mapping and Localization , 2000 .

[41]  Xing Chen,et al.  Foveated observation of shape and motion , 2003, 2003 IEEE International Conference on Robotics and Automation (Cat. No.03CH37422).

[42]  Markus Maurer,et al.  A compact vision system for road vehicle guidance , 1996, Proceedings of 13th International Conference on Pattern Recognition.

[43]  S. Schaal,et al.  Humanoid Oculomotor Control Based on Concepts of Computational Neuroscience , 2001 .

[44]  Takeshi Ohashi,et al.  Obstacle avoidance and path planning for humanoid robots using stereo vision , 2004, IEEE International Conference on Robotics and Automation, 2004. Proceedings. ICRA '04. 2004.

[45]  Simon Lacroix,et al.  High resolution terrain mapping using low attitude aerial stereo imagery , 2003, Proceedings Ninth IEEE International Conference on Computer Vision.

[46]  Paolo Pirjanian,et al.  The vSLAM Algorithm for Robust Localization and Mapping , 2005, Proceedings of the 2005 IEEE International Conference on Robotics and Automation.

[47]  Martin Buss,et al.  Combined Trajectory Planning and Gaze Direction Control for Robotic Exploration , 2007, Proceedings 2007 IEEE International Conference on Robotics and Automation.

[48]  Matthew R. Walter,et al.  Sparse extended information filters: insights into sparsification , 2005, 2005 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[49]  Gamini Dissanayake,et al.  An efficient multiple hypothesis filter for bearing-only SLAM , 2004, 2004 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (IEEE Cat. No.04CH37566).

[50]  Hugh F. Durrant-Whyte,et al.  Simultaneous Localization and Mapping with Sparse Extended Information Filters , 2004, Int. J. Robotics Res..

[51]  R. Brooks,et al.  The cog project: building a humanoid robot , 1999 .

[52]  Fadi Dornaika,et al.  Attentive Wide-Field Sensing for Visual Telepresence and Surveillance , 2004 .

[53]  Gamini Dissanayake,et al.  Multi-Step Look-Ahead Trajectory Planning in SLAM: Possibility and Necessity , 2005, Proceedings of the 2005 IEEE International Conference on Robotics and Automation.

[54]  Nikolaos Papanikolopoulos,et al.  Dual-camera system for multi-level activity recognition , 2004, 2004 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (IEEE Cat. No.04CH37566).