Vision Navigation Based on the Fusion of Feature-based Methods and Direct Methods

Vision navigation is an alternative to Global Position System (GPS) in environments where access to GPS is denied. Yet, the classical feature-based method has to extract feature observations from the image as input, which performs poor in environments with little features. To address this problem, a fusion of feature-based method and direct method is designed. The feature-based method is used in feature rich regions, while the direct method is used in regions with little features. Thus we utilize the fusion of two methods to improve the environmental adaptability of vision navigation. To improve the robustness to outliers, the Huber weight function is applied. Then the nonlinear optimization method is utilized to obtain the optimal camera pose. Experimental results demonstrate that the proposed method can meet the needs of real-time autonomous navigation.

[1]  Andrew W. Fitzgibbon,et al.  KinectFusion: Real-time dense surface mapping and tracking , 2011, 2011 10th IEEE International Symposium on Mixed and Augmented Reality.

[2]  Javier Civera,et al.  Using superpixels in monocular SLAM , 2014, 2014 IEEE International Conference on Robotics and Automation (ICRA).

[3]  Tom Drummond,et al.  Edge landmarks in monocular SLAM , 2009, Image Vis. Comput..

[4]  Daniel Cremers,et al.  LSD-SLAM: Large-Scale Direct Monocular SLAM , 2014, ECCV.

[5]  Daniel Cremers,et al.  Semi-dense Visual Odometry for a Monocular Camera , 2013, 2013 IEEE International Conference on Computer Vision.

[6]  Davide Scaramuzza,et al.  SVO: Fast semi-direct monocular visual odometry , 2014, 2014 IEEE International Conference on Robotics and Automation (ICRA).

[7]  Timothy D. Barfoot,et al.  State Estimation for Robotics , 2017 .

[8]  G. Klein,et al.  Parallel Tracking and Mapping for Small AR Workspaces , 2007, 2007 6th IEEE and ACM International Symposium on Mixed and Augmented Reality.

[9]  Kurt Konolige,et al.  Double window optimisation for constant time visual SLAM , 2011, 2011 International Conference on Computer Vision.

[10]  David W. Murray,et al.  Improving the Agility of Keyframe-Based SLAM , 2008, ECCV.

[11]  Hujun Bao,et al.  Robust monocular SLAM in dynamic environments , 2013, 2013 IEEE International Symposium on Mixed and Augmented Reality (ISMAR).

[12]  Jörg Stückler,et al.  Large-scale direct SLAM with stereo cameras , 2015, 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[13]  Wolfram Burgard,et al.  A benchmark for the evaluation of RGB-D SLAM systems , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[14]  Daniel Cremers,et al.  Semi-dense visual odometry for AR on a smartphone , 2014, 2014 IEEE International Symposium on Mixed and Augmented Reality (ISMAR).

[15]  J. M. M. Montiel,et al.  ORB-SLAM: A Versatile and Accurate Monocular SLAM System , 2015, IEEE Transactions on Robotics.

[16]  张振杰 Zhang Zhen-jie,et al.  Iteratively reweighted least squares method for camera pose estimation based on coplanar line correspondences , 2016 .

[17]  Andrew J. Davison,et al.  DTAM: Dense tracking and mapping in real-time , 2011, 2011 International Conference on Computer Vision.