Two years of Visual Odometry on the Mars Exploration Rovers

NASA's two Mars Exploration Rovers (MER) have successfully demonstrated a robotic Visual Odometry capability on another world for the first time. This provides each rover with accurate knowledge of its position, allowing it to autonomously detect and compensate for any unforeseen slip encountered during a drive. It has enabled the rovers to drive safely and more effectively in highly sloped and sandy terrains and has resulted in increased mission science return by reducing the number of days required to drive into interesting areas. The MER Visual Odometry system comprises onboard software for comparing stereo pairs taken by the pointable mast-mounted 45 deg FOV Navigation cameras (NAVCAMs). The system computes an update to the 6 degree of freedom rover pose (x, y, z, roll, pitch, yaw) by tracking the motion of autonomously selected terrain features between two pairs of 256×256 stereo images. It has demonstrated good performance with high rates of successful convergence (97% on Spirit, 95% on Opportunity), successfully detected slip ratios as high as 125%, and measured changes as small as 2 mm, even while driving on slopes as high as 31 deg. Visual Odometry was used over 14% of the first 10.7 km driven by both rovers. During the first 2 years of operations, Visual Odometry evolved from an “extra credit” capability into a critical vehicle safety system. In this paper we describe our Visual Odometry algorithm, discuss several driving strategies that rely on it (including Slip Checks, Keep-out Zones, and Wheel Dragging), and summarize its results from the first 2 years of operations on Mars. © 2006 Wiley Periodicals, Inc.

[1]  Clark F. Olson,et al.  Rover navigation using stereo ego-motion , 2003, Robotics Auton. Syst..

[2]  Saied Moezzi,et al.  Dynamic stereo vision , 1992 .

[3]  Alfred S. McEwen,et al.  Orogenic tectonism on Io , 2001 .

[4]  Takeo Kanade,et al.  A visual odometer for autonomous helicopter flight , 1999, Robotics Auton. Syst..

[5]  Simon Lacroix,et al.  Rover Self Localization in Planetary-Like Environments , 1999 .

[6]  Christopher G. Harris,et al.  A Combined Corner and Edge Detector , 1988, Alvey Vision Conference.

[7]  David Nistér,et al.  An efficient solution to the five-point relative pose problem , 2004, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[8]  M. Golombek,et al.  Size‐frequency distributions of rocks on Mars and Earth analog sites: Implications for future landed missions , 1997 .

[9]  P. Schönemann,et al.  Fitting one matrix to another under choice of a central dilation and a rigid motion , 1970 .

[10]  James J. Little,et al.  Mobile Robot Localization and Mapping with Uncertainty using Scale-Invariant Visual Landmarks , 2002, Int. J. Robotics Res..

[11]  Hans P. Moravec Obstacle avoidance and navigation in the real world by a seeing robot rover , 1980 .

[12]  M. Klimesh,et al.  Mars Exploration Rover engineering cameras , 2003 .

[13]  Shree K. Nayar,et al.  Ego-motion and omnidirectional cameras , 1998, Sixth International Conference on Computer Vision (IEEE Cat. No.98CH36271).

[14]  James R. Bergen,et al.  Visual odometry , 2004, Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2004. CVPR 2004..

[15]  Olivier D. Faugeras,et al.  Analysis Of A Sequence Of Stereo Scenes Containing Multiple Moving Objects Using Rigidity Constraints , 1988, [1988 Proceedings] Second International Conference on Computer Vision.

[16]  Edward Tunstel,et al.  Mars exploration rover surface operations: driving opportunity at Meridiani Planum , 2005, IEEE Robotics & Automation Magazine.

[17]  Clark F. Olson,et al.  Maximum likelihood rover localization by matching range maps , 1998, Proceedings. 1998 IEEE International Conference on Robotics and Automation (Cat. No.98CH36146).

[18]  David Nistér,et al.  Preemptive RANSAC for live structure and motion estimation , 2005, Machine Vision and Applications.

[19]  Quanan Zheng,et al.  Winter‐spring fronts in Taiwan Strait , 2006 .

[20]  Andrew E. Johnson,et al.  Opportunity rover localization and topographic mapping at the landing site of Meridiani Planum, Mars , 2007 .

[21]  J.J. Biesiadecki,et al.  The Mars Exploration Rover surface mobility flight software driving ambition , 2006, 2006 IEEE Aerospace Conference.

[22]  Larry H. Matthies,et al.  Error Modelling in Stereo Navigation , 1986, FJCC.

[23]  K. Di,et al.  Spirit rover localization and topographic mapping at the landing site of Gusev crater, Mars , 2006 .

[24]  Randel A. Lindemann,et al.  Mars Exploration Rover mobility assembly design, test and performance , 2005, 2005 IEEE International Conference on Systems, Man and Cybernetics.

[25]  Larry H. Matthies,et al.  Error modeling in stereo navigation , 1986, IEEE J. Robotics Autom..

[26]  Yang Cheng,et al.  Path following using visual odometry for a Mars rover in high-slip environments , 2004, 2004 IEEE Aerospace Conference Proceedings (IEEE Cat. No.04TH8720).

[27]  Nick Barnes,et al.  Performance of optical flow techniques for indoor navigation with a mobile robot , 2004, IEEE International Conference on Robotics and Automation, 2004. Proceedings. ICRA '04. 2004.

[28]  Peter I. Corke,et al.  Omnidirectional visual odometry for a planetary rover , 2004, 2004 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (IEEE Cat. No.04CH37566).

[29]  James R. Bergen,et al.  Visual odometry for ground vehicle applications , 2006, J. Field Robotics.

[30]  Jeffrey J. Biesiadecki,et al.  Tradeoffs Between Directed and Autonomous Driving on the Mars Exploration Rovers , 2007, ISRR.

[31]  Hans J. Schneebeli,et al.  A general approach for egomotion estimation with omnidirectional images , 2002, Proceedings of the IEEE Workshop on Omnidirectional Vision 2002. Held in conjunction with ECCV'02.