Discovering Planes and Collapsing the State Space in Visual SLAM

Recent advances in real-time visual SLAM have been based primarily on mapping isolated 3-D points. This presents difficulties when seeking to extend operation to wide areas, as the system state becomes large, requiring increasing computational effort. In this paper we present a novel approach to this problem in which planar structural components are embedded within the state to represent mapped points lying on a common plane. This collapses the state size, reducing computation and improving scalability, as well as giving a higher level scene description. Critically, the plane parameters are augmented into the SLAM state in a proper fashion, maintaining inherent uncertainties via a full covariance representation. Results for simulated data and for real-time operation demonstrate that the approach is effective.

[1]  David W. Murray,et al.  Interaction between hand and wearable camera in 2d and 3d environments , 2004, BMVC.

[2]  Tom Drummond,et al.  Edge landmarks in monocular SLAM , 2009, Image Vis. Comput..

[3]  Javier Civera,et al.  Inverse Depth to Depth Conversion for Monocular SLAM , 2007, Proceedings 2007 IEEE International Conference on Robotics and Automation.

[4]  Tom Drummond,et al.  Machine Learning for High-Speed Corner Detection , 2006, ECCV.

[5]  H. Durrant-Whyte,et al.  Simultaneous Localisation and Mapping ( SLAM ) : Part II State of the Art , 2006 .

[6]  Walterio W. Mayol-Cuevas,et al.  Real-Time Model-Based SLAM Using Line Segments , 2006, ISVC.

[7]  Ian D. Reid,et al.  Real-Time Monocular SLAM with Straight Lines , 2006, BMVC.

[8]  Manolis I. A. Lourakis,et al.  Estimating the Jacobian of the Singular Value Decomposition: Theory and Applications , 2000, ECCV.

[9]  Ian D. Reid,et al.  Locally Planar Patch Features for Real-Time Structure from Motion , 2004, BMVC.

[10]  Andrew J. Davison,et al.  Real-time simultaneous localisation and mapping with a single camera , 2003, Proceedings Ninth IEEE International Conference on Computer Vision.

[11]  Javier Civera,et al.  Unified Inverse Depth Parametrization for Monocular SLAM , 2006, Robotics: Science and Systems.

[12]  Hugh Durrant-Whyte,et al.  Simultaneous localization and mapping (SLAM): part II , 2006 .

[13]  Walterio W. Mayol-Cuevas,et al.  Robust Real-Time Visual SLAM Using Scale Prediction and Exemplar Based Feature Description , 2007, 2007 IEEE Conference on Computer Vision and Pattern Recognition.

[14]  Ian D. Reid,et al.  Automatic Relocalisation for a Single-Camera Simultaneous Localisation and Mapping System , 2007, Proceedings 2007 IEEE International Conference on Robotics and Automation.

[15]  Peter Cheeseman,et al.  A stochastic map for uncertain spatial relationships , 1988 .

[16]  Adrien Bartoli,et al.  A random sampling strategy for piecewise planar scene segmentation , 2007, Comput. Vis. Image Underst..

[17]  Tom Drummond,et al.  Scalable Monocular SLAM , 2006, 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06).

[18]  Robert C. Bolles,et al.  Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography , 1981, CACM.

[19]  Hugh Durrant-Whyte,et al.  Simultaneous Localisation and Mapping ( SLAM ) : Part I The Essential Algorithms , 2006 .

[20]  Thia Kirubarajan,et al.  Estimation with Applications to Tracking and Navigation: Theory, Algorithms and Software , 2001 .

[21]  David W. Murray,et al.  Towards simultaneous recognition, localization and mapping for hand-held and wearable cameras , 2007, Proceedings 2007 IEEE International Conference on Robotics and Automation.