Robust Appearance Based Visual Route Following for Navigation in Large-scale Outdoor Environments

In this paper we present a navigation algorithm that enables mobile robots to retrace routes previously taught under the control of human operators in outdoor environments. Possible applications include robot couriers, autonomous vehicles, tour guides and robotic patrols. The appearance-based approach presented in the paper is provably convergent, computationally inexpensive compared with map-based approaches and requires only odometry and a monocular omnidirectional vision sensor. A sequence of reference images is recorded during the human-guided route-teaching phase. Before starting the autonomous phase, the robot needs to be positioned at the beginning of the route. During the autonomous phase, the measurement image is compared with reference images using image cross-correlation performed in the Fourier domain to recover the difference in relative orientation. Route following is achieved by compensating for this orientation difference. Over 18 km of experiments performed under varying conditions demonstrate the algorithm's robustness to lighting variations and partial occlusion. Obstacle avoidance is not included in the current system.

[1]  Masayuki Inaba,et al.  View-based approach to robot navigation , 2000, Proceedings. 2000 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2000) (Cat. No.00CH37113).

[2]  M. Srinivasan Animal navigation: Ants match as they march , 1998, Nature.

[3]  Emanuele Menegatti,et al.  Image-based memory for robot navigation using properties of omnidirectional images , 2004, Robotics Auton. Syst..

[4]  Andrew Vardy,et al.  Low-Level Visual Homing , 2003, ECAL.

[5]  Bernhard Schölkopf,et al.  Where did I take that snapshot? Scene-based homing by image matching , 1998, Biological Cybernetics.

[6]  Kostas Daniilidis,et al.  Rotation recovery from spherical images without correspondences , 2006, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[7]  Frédéric Labrosse,et al.  The visual compass: Performance and limitations of an appearance‐based method , 2006, J. Field Robotics.

[8]  Ales Leonardis,et al.  Panoramic volumes for robot localization , 2005, 2005 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[9]  Masayuki Inaba,et al.  Visual navigation using omnidirectional view sequence , 1999, Proceedings 1999 IEEE/RSJ International Conference on Intelligent Robots and Systems. Human and Environment Friendly Robots with High Intelligence and Emotional Quotients (Cat. No.99CH36289).

[10]  Ben J. A. Kröse,et al.  Navigation using an appearance based topological map , 2007, Proceedings 2007 IEEE International Conference on Robotics and Automation.

[11]  M. Srinivasan,et al.  A complete panoramic vision system, incorporating imaging, ranging, and three dimensional navigation , 2000, Proceedings IEEE Workshop on Omnidirectional Vision (Cat. No.PR00704).

[12]  Gamini Dissanayake,et al.  Simultaneous Localisation and Mapping: A Stereo Vision Based Approach , 2006, 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[13]  Luc Van Gool,et al.  Feature based omnidirectional sparse visual path following , 2005, 2005 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[14]  Sinisa Segvic,et al.  Large scale vision-based navigation without an accurate global reconstruction , 2007, 2007 IEEE Conference on Computer Vision and Pattern Recognition.

[15]  Thomas S. Collett,et al.  Landmark learning and guidance in insects , 1992 .

[16]  Antonis A. Argyros,et al.  Robot homing based on corner tracking in a sequence of panoramic images , 2001, Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition. CVPR 2001.

[17]  Frédéric Labrosse Short and long-range visual navigation using warped panoramic images , 2007, Robotics Auton. Syst..

[18]  Naokazu Yokoya,et al.  Memory-based self-localization using omnidirectional images , 1998, Proceedings. Fourteenth International Conference on Pattern Recognition (Cat. No.98EX170).

[19]  Ales Leonardis,et al.  Robust localization using an omnidirectional appearance-based subspace model of environment , 2003, Robotics Auton. Syst..

[20]  Ian D. Reid,et al.  Real-Time Monocular SLAM with Straight Lines , 2006, BMVC.

[21]  Hanspeter A. Mallot,et al.  Efficient visual homing based on Fourier transformed panoramic images , 2006, Robotics Auton. Syst..

[22]  James L. Crowley,et al.  Appearance based processes for visual navigation , 1997 .

[23]  Michel Dhome,et al.  Outdoor autonomous navigation using monocular vision , 2005, 2005 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[24]  Lixin Tang,et al.  Indoor navigation for mobile robots using memorized omni-directional images and robot's motion , 2002, IEEE/RSJ International Conference on Intelligent Robots and Systems.

[25]  Václav Hlavác,et al.  Zero Phase Representation of Panoramic Images for Image Vased Localization , 1999, CAIP.

[26]  T. Collett,et al.  Multiple stored views and landmark guidance in ants , 1998, Nature.

[27]  Masayuki Inaba,et al.  Visual navigation using view-sequenced route representation , 1996, Proceedings of IEEE International Conference on Robotics and Automation.

[28]  Tom,et al.  Epipolar Geometry for Panoramic Cameras Epipolar Geometry for Panoramic Cameras ? , 1998 .

[29]  W. Burgard,et al.  Markov Localization for Mobile Robots in Dynamic Environments , 1999, J. Artif. Intell. Res..

[30]  Andrew J. Davison,et al.  Real-time simultaneous localisation and mapping with a single camera , 2003, Proceedings Ninth IEEE International Conference on Computer Vision.

[31]  Yasushi Yagi,et al.  Making 2D Map of Environments Based upon Routes Scenes , 2000, Auton. Robots.