Gaussian Process Gauss-Newton for 3D laser-based Visual Odometry

In this paper, we present a method for obtaining Visual Odometry (VO) estimates using a scanning laser rangefinder. Though common VO implementations utilize stereo camera imagery, cameras are dependent on ambient light. In contrast, actively-illuminated sensors such as laser rangefinders work in a variety of lighting conditions, including full darkness. We leverage previous successes by applying sparse appearance-based methods to laser intensity images, and address the issue of motion distortion by considering the estimation problem in continuous time. This is facilitated by Gaussian Process Gauss-Newton (GPGN), an algorithm for non-parametric, continuous-time, nonlinear, batch state estimation. We include a concise derivation of GPGN, along with details on the extension to three-dimensions (3D). Validation of the 3D laser-based VO framework is provided using 1.1km of experimental data, which was gathered by a field robot equipped with a two-axis scanning lidar.

[1]  Ken Shoemake,et al.  Animating rotation with quaternion curves , 1985, SIGGRAPH.

[2]  Gaurav S. Sukhatme,et al.  Visual-Inertial Sensor Fusion: Localization, Mapping and Sensor-to-Sensor Self-calibration , 2011, Int. J. Robotics Res..

[3]  Robert C. Bolles,et al.  Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography , 1981, CACM.

[4]  Christopher G. Harris,et al.  A Combined Corner and Edge Detector , 1988, Alvey Vision Conference.

[5]  Tim D. Barfoot,et al.  Towards relative continuous-time SLAM , 2013, 2013 IEEE International Conference on Robotics and Automation.

[6]  Per-Erik Forssén,et al.  Efficient Video Rectification and Stabilisation for Cell-Phones , 2012, International Journal of Computer Vision.

[7]  Christopher Hunt,et al.  Notes on the OpenSURF Library , 2009 .

[8]  John Stillwell,et al.  Naive Lie Theory , 2008 .

[9]  Gaurav S. Sukhatme,et al.  Sliding window filter with application to planetary landing , 2010, J. Field Robotics.

[10]  Ian D. Reid,et al.  A hybrid SLAM representation for dynamic marine environments , 2010, 2010 IEEE International Conference on Robotics and Automation.

[11]  Ian D. Reid,et al.  Vast-scale Outdoor Navigation Using Adaptive Relative Bundle Adjustment , 2010, Int. J. Robotics Res..

[12]  Peter J. Huber,et al.  Robust Statistics , 2005, Wiley Series in Probability and Statistics.

[13]  Larry H. Matthies,et al.  Two years of Visual Odometry on the Mars Exploration Rovers , 2007, J. Field Robotics.

[14]  Gabe Sibley,et al.  Sliding window filter with application to planetary landing , 2010 .

[15]  Michael Bosse,et al.  Continuous 3D scan-matching with a spinning 2D laser , 2009, 2009 IEEE International Conference on Robotics and Automation.

[16]  Luc Van Gool,et al.  Speeded-Up Robust Features (SURF) , 2008, Comput. Vis. Image Underst..

[17]  Carl E. Rasmussen,et al.  Gaussian processes for machine learning , 2005, Adaptive computation and machine learning.

[18]  Paul Timothy Furgale,et al.  Visual Teach and Repeat using appearance-based lidar , 2011, 2012 IEEE International Conference on Robotics and Automation.

[19]  Paul Timothy Furgale,et al.  Gaussian Process Gauss-Newton: Non-Parametric State Estimation , 2012, 2012 Ninth Conference on Computer and Robot Vision.

[20]  Michael Bosse,et al.  Zebedee: Design of a Spring-Mounted 3-D Range Sensor with Application to Mobile Mapping , 2012, IEEE Transactions on Robotics.

[21]  Hang Dong,et al.  Two-axis scanning lidar geometric calibration using intensity imagery and distortion mapping , 2013, 2013 IEEE International Conference on Robotics and Automation.

[22]  David G. Lowe,et al.  Distinctive Image Features from Scale-Invariant Keypoints , 2004, International Journal of Computer Vision.

[23]  Paul Timothy Furgale,et al.  Gaussian Process Gauss–Newton for non-parametric simultaneous localization and mapping , 2013, Int. J. Robotics Res..

[24]  R. E. Kalman,et al.  New Results in Linear Filtering and Prediction Theory , 1961 .

[25]  Paul Timothy Furgale,et al.  Continuous-time batch estimation using temporal basis functions , 2012, 2012 IEEE International Conference on Robotics and Automation.

[26]  G LoweDavid,et al.  Distinctive Image Features from Scale-Invariant Keypoints , 2004 .

[27]  Eduardo Mario Nebot,et al.  Localization and map building using laser range sensors in outdoor applications , 2000, J. Field Robotics.

[28]  Paul Timothy Furgale,et al.  Towards appearance-based methods for lidar sensors , 2011, 2011 IEEE International Conference on Robotics and Automation.

[29]  Joachim Hertzberg,et al.  6D SLAM—3D mapping outdoor environments , 2007, J. Field Robotics.

[30]  Günther Schmidt,et al.  Fusing range and intensity images for mobile robot localization , 1999, IEEE Trans. Robotics Autom..

[31]  B. Ripley,et al.  Robust Statistics , 2018, Encyclopedia of Mathematical Geosciences.

[32]  Kurt Konolige,et al.  Large-Scale Visual Odometry for Rough Terrain , 2007, ISRR.

[33]  Tim D. Barfoot,et al.  A self-calibrating 3D ground-truth localization system using retroreflective landmarks , 2011, 2011 IEEE International Conference on Robotics and Automation.

[34]  Hang Dong,et al.  Lighting-Invariant Visual Odometry using Lidar Intensity Imagery and Pose Interpolation , 2012, FSR.

[35]  Michael Bosse,et al.  Efficient Large-Scale 3D Mobile Mapping and Surface Reconstruction of an Underground Mine , 2012, FSR.