Single crater-aided inertial navigation for autonomous asteroid landing

Abstract In this paper, a novel crater-aided inertial navigation approach for autonomous asteroid landing mission is developed. It overcomes the major deficiencies of existing approaches in the literature, which mainly focuses on the case where craters are abundant in the camera field of view. As a result, traditional crater based methods require at least three craters to achieve crater matching, which limits their application in final landing phase where craters are scarce in the camera’s field of view. In contrast, the proposed algorithm enables single crater based crater matching based on a novel 2D-3D crater re-projection model. The re-projection model adopts inertial measurements as a reference, and re-projects the 3D crater model onto descent images to achieve the matching to its counterpart. An asteroid landing simulation toolbox is developed to validate the performance of the proposed approach. Through comparison with the state-of-the-art local image feature and crater based navigation algorithms, the proposed approach is validated to achieve a competitive performance in terms of feature matching and pose estimation accuracy with a much lighter computational cost.

[1]  John F. Canny,et al.  A Computational Approach to Edge Detection , 1986, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[2]  A. Konopliv,et al.  Modeling of the Eros gravity field as an ellipsoidal harmonic expansion from the NEAR Doppler tracking data , 2002 .

[3]  Roberto Furfaro,et al.  Asteroid Precision Landing via Multiple Sliding Surfaces Guidance Techniques , 2013 .

[4]  Wei Li,et al.  A machine learning approach to crater detection from topographic data , 2014 .

[5]  Keiken Ninomiya,et al.  Optical guidance for autonomous landing of spacecraft , 1999 .

[6]  Xiangyu Huang,et al.  Landmark-based autonomous navigation for pinpoint planetary landing , 2016 .

[7]  A.E. Johnson,et al.  Overview of Terrain Relative Navigation Approaches for Precise Lunar Landing , 2008, 2008 IEEE Aerospace Conference.

[8]  Wei Ding,et al.  Detection of Sub-Kilometer Craters in High Resolution Planetary Images Using Shape and Texture Features , 2012 .

[9]  Jinjun Shan,et al.  Reliability-Based Soft Landing Trajectory Optimization near Asteroid with Uncertain Gravitational Field , 2015 .

[10]  Francisco Bonin-Font,et al.  Visual Navigation for Mobile Robots: A Survey , 2008, J. Intell. Robotic Syst..

[11]  N. Izenberg,et al.  The landing of the NEAR-Shoemaker spacecraft on asteroid 433 Eros , 2001, Nature.

[12]  Luc Van Gool,et al.  SURF: Speeded Up Robust Features , 2006, ECCV.

[13]  Hutao Cui,et al.  A new approach based on crater detection and matching for visual navigation in planetary landing , 2014 .

[14]  Stergios I. Roumeliotis,et al.  Vision‐aided inertial navigation for pin‐point landing using observations of mapped landmarks , 2007, J. Field Robotics.

[15]  Shengying Zhu,et al.  Visual navigation using edge curve matching for pinpoint planetary landing , 2018 .

[16]  Andrew F. Cheng,et al.  Small-Scale Topography of 433 Eros from Laser Altimetry and Imaging , 2000 .

[17]  Gaurav S. Sukhatme,et al.  Vision-based autonomous landing of an unmanned aerial vehicle , 2002, Proceedings 2002 IEEE International Conference on Robotics and Automation (Cat. No.02CH37292).

[18]  Sven Loncaric,et al.  Open framework for objective evaluation of crater detection algorithms with first test-field subsystem based on MOLA data , 2008 .

[19]  Christopher G. Harris,et al.  A Combined Corner and Edge Detector , 1988, Alvey Vision Conference.

[20]  Michèle Lavagna,et al.  A multilayer perceptron hazard detector for vision-based autonomous planetary landing , 2016 .

[21]  Manolis I. A. Lourakis,et al.  High-Performance Embedded Computing in Space: Evaluation of Platforms for Vision-Based Navigation , 2018 .

[22]  J. Junkins,et al.  Analytical Mechanics of Space Systems , 2003 .

[23]  Larry H. Matthies,et al.  Crater detection for autonomous landing on asteroids , 2001, Image Vis. Comput..

[24]  Larry H. Matthies,et al.  Vision Guided Landing of an Autonomous Helicopter in Hazardous Terrain , 2005, Proceedings of the 2005 IEEE International Conference on Robotics and Automation.

[25]  Meng Ding,et al.  Removing dust impact for visual navigation in Mars landing , 2016 .

[26]  Pedro Pina,et al.  Automatic Recognition of Impact Craters on the Surface of Mars , 2004, ICIAR.

[27]  Stergios I. Roumeliotis,et al.  Vision-Aided Inertial Navigation for Spacecraft Entry, Descent, and Landing , 2009, IEEE Transactions on Robotics.

[28]  Horst Bischof,et al.  Efficient Maximally Stable Extremal Region (MSER) Tracking , 2006, 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06).

[29]  Hutao Cui,et al.  Database construction for vision aided navigation in planetary landing , 2017 .

[30]  Marco Piras,et al.  Performances comparison of different MEMS-based IMUs , 2010, IEEE/ION Position, Location and Navigation Symposium.

[31]  Clark F. Olson,et al.  Optical landmark detection for spacecraft navigation , 2003 .

[32]  David G. Lowe,et al.  Distinctive Image Features from Scale-Invariant Keypoints , 2004, International Journal of Computer Vision.

[33]  S. Govindarajulu,et al.  A Comparison of SIFT, PCA-SIFT and SURF , 2012 .

[34]  Xiangyu Huang,et al.  Radio beacons/IMU integrated navigation for Mars entry , 2011 .

[35]  J. Betts Survey of Numerical Methods for Trajectory Optimization , 1998 .