Dense Localization of a Monocular Camera Using Keyframes

In this paper, we present a low cost localization system that exploits dense image information to continuously track the position of a camera in 6DOF. It leverages of the use of a set of selected "key frames" separated in distance from which a depth map is available to create a local 3D point cloud. In this way, we avoid the computational overload caused by common dense sequential approaches. The system uses a 3D-2D technique to calculate an initial pose estimate for the intermediate camera frames. A refinement step stated as a Non Linear Least Squares (NLQs) optimisation is performed by minimising the photo-consistency error. The NLQs cost function is defined by aligning a warped image and an image associated to the closest key frame. The minimum solution is calculated using the Levenberg-Marquardt method. To validate the accuracy of our system, we conducted experiments using data with perfect ground truth. Our assessment shows that our system is able to achieve up to millimeter accuracy. Most of the expensive calculations are carried out by exploiting parallel computing and GPGPU.

[1]  Andrew J. Davison,et al.  DTAM: Dense tracking and mapping in real-time , 2011, 2011 International Conference on Computer Vision.

[2]  Andrew J. Davison,et al.  Real-Time Camera Tracking: When is High Frame-Rate Best? , 2012, ECCV.

[3]  Simon Baker,et al.  Lucas-Kanade 20 Years On: A Unifying Framework , 2004, International Journal of Computer Vision.

[4]  David W. Murray,et al.  Improving the Agility of Keyframe-Based SLAM , 2008, ECCV.

[5]  G. Bellamy Lie groups, Lie algebras, and their representations , 2015 .

[6]  Patrick Rives,et al.  An Efficient Direct Method for Improving visual SLAM , 2007, Proceedings 2007 IEEE International Conference on Robotics and Automation.

[7]  G. Klein,et al.  Parallel Tracking and Mapping for Small AR Workspaces , 2007, 2007 6th IEEE and ACM International Symposium on Mixed and Augmented Reality.

[8]  Andrew J. Davison,et al.  Real-time simultaneous localisation and mapping with a single camera , 2003, Proceedings Ninth IEEE International Conference on Computer Vision.

[9]  Javier Civera,et al.  Inverse Depth Parametrization for Monocular SLAM , 2008, IEEE Transactions on Robotics.

[10]  Andrew J. Davison,et al.  Live dense reconstruction with a single moving camera , 2010, 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[11]  Andrew J. Davison,et al.  Real-Time Spherical Mosaicing Using Whole Image Alignment , 2010, ECCV.

[12]  Andrew Zisserman,et al.  Multiple View Geometry in Computer Vision (2nd ed) , 2003 .

[13]  Daniel Cremers,et al.  LSD-SLAM: Large-Scale Direct Monocular SLAM , 2014, ECCV.

[14]  Daniel Cremers,et al.  Real-Time Dense Geometry from a Handheld Camera , 2010, DAGM-Symposium.

[15]  Daniel Cremers,et al.  Real-time visual odometry from dense RGB-D images , 2011, 2011 IEEE International Conference on Computer Vision Workshops (ICCV Workshops).

[16]  C. Zach Fast and High Quality Fusion of Depth Maps , 2008 .

[17]  Lina María Paz,et al.  Divide and Conquer: EKF SLAM in O(n) , 2008, IEEE Trans. Robotics.

[18]  Olivier Stasse,et al.  MonoSLAM: Real-Time Single Camera SLAM , 2007, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[19]  David Nister,et al.  Recent developments on direct relative orientation , 2006 .

[20]  Bernhard P. Wrobel,et al.  Multiple View Geometry in Computer Vision , 2001 .