UV-SLAM: Unconstrained Line-Based SLAM Using Vanishing Points for Structural Mapping

In feature-based simultaneous localization and mapping (SLAM), line features complement the sparsity of point features, making it possible to map the surrounding environment structure. Existing approaches utilizing line features have primarily employed a measurement model that uses line re-projection. However, the direction vectors used in the 3D line mapping process cannot be corrected because the line measurement model employs only the lines’ normal vectors in the Plücker coordinate. As a result, problems like degeneracy that occur during the 3D line mapping process cannot be solved. To tackle the problem, this paper presents a UV-SLAM, which is an unconstrained line-based SLAM using vanishing points for structural mapping. This paper focuses on using structural regularities without any constraints, such as the Manhattan world assumption. For this, we use the vanishing points that can be obtained from the line features. The difference between the vanishing point observation calculated through line features in the image and the vanishing point estimation calculated through the direction vector is defined as a residual and added to the cost function of optimization-based SLAM. Furthermore, through Fisher information matrix rank analysis, we prove that vanishing point measurements guarantee a unique mapping solution. Finally, we demonstrate that the localization accuracy and mapping quality are improved compared to the state-ofthe-art algorithms using public datasets.

[1]  Yong Liu,et al.  Robust visual SLAM with point and line features , 2017, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[2]  Francesc Moreno-Noguer,et al.  PL-SLAM: Real-time monocular visual SLAM with points and lines , 2017, 2017 IEEE International Conference on Robotics and Automation (ICRA).

[3]  Hyun Myung,et al.  Avoiding Degeneracy for Monocular Visual SLAM with Point and Line Features , 2021, 2021 IEEE International Conference on Robotics and Automation (ICRA).

[4]  Gamini Dissanayake,et al.  Observability analysis of SLAM using fisher information matrix , 2008, 2008 10th International Conference on Control, Automation, Robotics and Vision.

[5]  Nikolas Brasch,et al.  Structure-SLAM: Low-Drift Monocular SLAM in Indoor Environments , 2020, IEEE Robotics and Automation Letters.

[6]  Sang Jun Lee,et al.  Elaborate Monocular Point and Line SLAM With Robust Initialization , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).

[7]  Gaurav S. Sukhatme,et al.  Sliding window filter with application to planetary landing , 2010, J. Field Robotics.

[8]  Nikolas Brasch,et al.  RGB-D SLAM with Structural Regularities , 2021, 2021 IEEE International Conference on Robotics and Automation (ICRA).

[9]  Brian Coltin,et al.  Indoor RGB-D Compass from a Single Line and Plane , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[10]  Yuanxin Wu,et al.  StructVIO: Visual-Inertial Odometry With Structural Regularity of Man-Made Environments , 2018, IEEE Transactions on Robotics.

[11]  Andrea Fusiello,et al.  Robust Multiple Structures Estimation with J-Linkage , 2008, ECCV.

[12]  Rafael Grompone von Gioi,et al.  LSD: A Fast Line Segment Detector with a False Detection Control , 2010, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[13]  Feng Guo,et al.  PL-VINS: Real-Time Monocular Visual-Inertial SLAM with Point and Line , 2020, ArXiv.

[14]  Frederick R. Forst,et al.  On robust estimation of the location parameter , 1980 .

[15]  Thia Kirubarajan,et al.  Estimation with Applications to Tracking and Navigation: Theory, Algorithms and Software , 2001 .

[16]  Zhe Zhang,et al.  Trifo-VIO: Robust and Efficient Stereo Visual Inertial Odometry Using Points and Lines , 2018, 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[17]  H. Jin Kim,et al.  Low-Drift Visual Odometry in Structured Environments by Decoupling Rotational and Translational Motion , 2018, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[18]  Yunhui Liu,et al.  SLAM-based 3D Line Reconstruction , 2018, 2018 13th World Congress on Intelligent Control and Automation (WCICA).

[19]  Ji Zhao,et al.  PL-VIO: Tightly-Coupled Monocular Visual–Inertial Odometry Using Point and Line Features , 2018, Sensors.

[20]  Xiaoguang Mei,et al.  Line-Based Stereo SLAM by Junction Matching and Vanishing Point Alignment , 2019, IEEE Access.

[21]  Masatoshi Okutomi,et al.  3D Surface Reconstruction from Point-and-Line Cloud , 2015, 2015 International Conference on 3D Vision.

[22]  Soon-Yong Park,et al.  PLF-VINS: Real-Time Monocular Visual-Inertial SLAM With Point-Line Fusion and Parallel-Line Fusion , 2021, IEEE Robotics and Automation Letters.

[23]  Christian Heipke,et al.  Accurate Reconstruction of Near-Epipolar Line Segments from Stereo Aerial Images , 2012 .

[24]  Shaojie Shen,et al.  VINS-Mono: A Robust and Versatile Monocular Visual-Inertial State Estimator , 2017, IEEE Transactions on Robotics.

[25]  Wenqi Wu,et al.  Tightly-Coupled Stereo Visual-Inertial Navigation Using Point and Line Features , 2015, Sensors.

[26]  J. M. M. Montiel,et al.  ORB-SLAM: A Versatile and Accurate Monocular SLAM System , 2015, IEEE Transactions on Robotics.

[27]  Roland Siegwart,et al.  The EuRoC micro aerial vehicle datasets , 2016, Int. J. Robotics Res..

[28]  Guoquan Huang,et al.  Visual-Inertial Odometry with Point and Line Features , 2019, 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[29]  Hyun Myung,et al.  ALVIO: Adaptive Line and Point Feature-based Visual Inertial Odometry for Robust Localization in Indoor Environments , 2020, ArXiv.

[30]  Federico Tombari,et al.  ManhattanSLAM: Robust Planar Tracking and Mapping Leveraging Mixture of Manhattan Frames , 2021, 2021 IEEE International Conference on Robotics and Automation (ICRA).

[31]  Il Hong Suh,et al.  Building a 3-D Line-Based Map Using Stereo SLAM , 2015, IEEE Transactions on Robotics.

[32]  Stergios I. Roumeliotis,et al.  A Multi-State Constraint Kalman Filter for Vision-aided Inertial Navigation , 2007, Proceedings 2007 IEEE International Conference on Robotics and Automation.

[33]  Bernhard P. Wrobel,et al.  Multiple View Geometry in Computer Vision , 2001 .

[34]  Reinhard Koch,et al.  An efficient and robust line segment matching approach based on LBD descriptor and pairwise geometric consistency , 2013, J. Vis. Commun. Image Represent..

[35]  Davide Scaramuzza,et al.  A Tutorial on Quantitative Trajectory Evaluation for Visual(-Inertial) Odometry , 2018, 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[36]  Dimitrios G. Kottas,et al.  Efficient and consistent vision-aided inertial navigation using line observations , 2013, 2013 IEEE International Conference on Robotics and Automation.

[37]  Frank Dellaert,et al.  On-Manifold Preintegration for Real-Time Visual--Inertial Odometry , 2015, IEEE Transactions on Robotics.

[38]  Guoquan Huang,et al.  Observability Analysis of Aided INS With Heterogeneous Features of Points, Lines, and Planes , 2018, IEEE Transactions on Robotics.

[39]  P. J. Huber Robust Estimation of a Location Parameter , 1964 .

[40]  Peng Wang,et al.  Leveraging Structural Information to Improve Point Line Visual-Inertial Odometry , 2021, IEEE Robotics and Automation Letters.

[41]  Danping Zou,et al.  StructSLAM: Visual SLAM With Building Structure Lines , 2015, IEEE Transactions on Vehicular Technology.

[42]  Adrien Bartoli,et al.  Structure-from-motion using lines: Representation, triangulation, and bundle adjustment , 2005, Comput. Vis. Image Underst..

[43]  Carlo Tomasi,et al.  Good features to track , 1994, 1994 Proceedings of IEEE Conference on Computer Vision and Pattern Recognition.