A Stereo Visual-Inertial SLAM Approach for Indoor Mobile Robots in Unknown Environments Without Occlusions

When mobile robots are working in indoor unknown environments, the surrounding scenes are mainly low texture or repeating texture. This means that image features are easily lost when tracking the robots, and poses are difficult to estimate as the robot moves back and forth in a narrow area. In order to improve such tracking problems, we propose a one-circle feature-matching method, which refers to a sequence of the circle matching for the time after space (STCM), and an STCM-based visual-inertial simultaneous localization and mapping (STCM-SLAM) technique. This strategy tightly couples the stereo camera and the inertial measurement unit (IMU) in order to better estimate poses of the mobile robot when working indoors. Forward backward optical flow is used to track image features. The absolute accuracy and relative accuracy of STCM increase by 37.869% and 129.167%, respectively, when compared with correlation flow. In addition, we compare our proposed method with other state-of-the-art methods. In terms of relative pose error, the accuracy of STCM-SLAM is an order of magnitude greater than ORB-SLAM2, and two orders of magnitude greater than OKVIS. Our experiments show that STCM-SLAM has obvious advantages over the OKVIS method, specifically in terms of scale error, running frequency, and CPU load. STCM-SLAM also performs the best under real-time conditions. In the indoor experiments, STCM-SLAM is able to accurately estimate the trajectory of the mobile robot. Based on the root mean square error, mean error, and standard deviation, the accuracy of STCM-SLAM is ultimately superior to that of either ORB-SLAM2 or OKVIS.

[1]  Michael Kaess,et al.  Generic Node Removal for Factor-Graph SLAM , 2014, IEEE Transactions on Robotics.

[2]  Rafael Muñoz-Salinas,et al.  UcoSLAM: Simultaneous Localization and Mapping by Fusion of KeyPoints and Squared Planar Markers , 2019, Pattern Recognit..

[3]  Zhaoxiang Liu,et al.  A Novel Feedback Mechanism-Based Stereo Visual-Inertial SLAM , 2019, IEEE Access.

[4]  Juan D. Tardós,et al.  Visual-Inertial Monocular SLAM With Map Reuse , 2016, IEEE Robotics and Automation Letters.

[5]  Ivan Markovic,et al.  SOFT‐SLAM: Computationally efficient stereo visual simultaneous localization and mapping for autonomous unmanned aerial vehicles , 2018, J. Field Robotics.

[6]  Hua Zhu,et al.  A Review of Visual-Inertial Simultaneous Localization and Mapping from Filtering-Based and Optimization-Based Perspectives , 2018, Robotics.

[7]  Mingyang Li,et al.  Improving the accuracy of EKF-based visual-inertial odometry , 2012, 2012 IEEE International Conference on Robotics and Automation.

[8]  Michael Bosse,et al.  Keyframe-based visual–inertial odometry using nonlinear optimization , 2015, Int. J. Robotics Res..

[9]  Javier González,et al.  Learning-Based Image Enhancement for Visual Odometry in Challenging HDR Environments , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[10]  Roland Siegwart,et al.  The EuRoC micro aerial vehicle datasets , 2016, Int. J. Robotics Res..

[11]  Chen Wang,et al.  Correlation Flow: Robust Optical Flow Using Kernel Cross-Correlators , 2018, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[12]  Rafael Muñoz-Salinas,et al.  SPM-SLAM: Simultaneous localization and mapping with squared planar markers , 2019, Pattern Recognit..

[13]  Tom Drummond,et al.  Machine Learning for High-Speed Corner Detection , 2006, ECCV.

[14]  Roland Siegwart,et al.  Robust visual inertial odometry using a direct EKF-based approach , 2015, IROS 2015.

[15]  Dongbing Gu,et al.  UnDeepVO: Monocular Visual Odometry Through Unsupervised Deep Learning , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[16]  Andrew J. Davison,et al.  DTAM: Dense tracking and mapping in real-time , 2011, 2011 International Conference on Computer Vision.

[17]  Artur Sagitov,et al.  Comparing Fiducial Markers Performance for a Task of a Humanoid Robot Self-calibration of Manipulators: A Pilot Experimental Study , 2018, ICR.

[18]  Stefan Leutenegger,et al.  ElasticFusion: Real-time dense SLAM and light source estimation , 2016, Int. J. Robotics Res..

[19]  G. Klein,et al.  Parallel Tracking and Mapping for Small AR Workspaces , 2007, 2007 6th IEEE and ACM International Symposium on Mixed and Augmented Reality.

[20]  Daniel Cremers,et al.  Dense visual SLAM for RGB-D cameras , 2013, 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[21]  Juan D. Tardós,et al.  ORB-SLAM2: An Open-Source SLAM System for Monocular, Stereo, and RGB-D Cameras , 2016, IEEE Transactions on Robotics.

[22]  Davide Scaramuzza,et al.  A Tutorial on Quantitative Trajectory Evaluation for Visual(-Inertial) Odometry , 2018, 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[23]  Ji Zhang,et al.  Visual-lidar odometry and mapping: low-drift, robust, and fast , 2015, 2015 IEEE International Conference on Robotics and Automation (ICRA).

[24]  Cordelia Schmid,et al.  Scale & Affine Invariant Interest Point Detectors , 2004, International Journal of Computer Vision.

[25]  Rafael Muñoz-Salinas,et al.  Simultaneous Multi-View Camera Pose Estimation and Object Tracking With Squared Planar Markers , 2019, IEEE Access.

[26]  Hugh Durrant-Whyte,et al.  Simultaneous localization and mapping (SLAM): part II , 2006 .

[27]  Hujun Bao,et al.  ICE-BA: Incremental, Consistent and Efficient Bundle Adjustment for Visual-Inertial SLAM , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[28]  Juan Song,et al.  Semantic SLAM Based on Object Detection and Improved Octomap , 2018, IEEE Access.

[29]  Kostas Daniilidis,et al.  TagSLAM: Robust SLAM with Fiducial Markers , 2019, ArXiv.

[30]  Lei Guo,et al.  An Automatic Key-Frame Selection Method for Monocular Visual Odometry of Ground Vehicle , 2019, IEEE Access.

[31]  Jiamao Li,et al.  Stereo Visual-Inertial SLAM With Points and Lines , 2018, IEEE Access.

[32]  Danping Zou,et al.  CoSLAM: Collaborative Visual SLAM in Dynamic Environments , 2013, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[33]  Tomasz Malisiewicz,et al.  SuperPoint: Self-Supervised Interest Point Detection and Description , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).

[34]  Hua Zhu,et al.  Visual-inertial SLAM method based on optical flow in a GPS-denied environment , 2018, Ind. Robot.

[35]  Shaojie Shen,et al.  VINS-Mono: A Robust and Versatile Monocular Visual-Inertial State Estimator , 2017, IEEE Transactions on Robotics.

[36]  Frank Dellaert,et al.  iSAM2: Incremental smoothing and mapping with fluid relinearization and incremental variable reordering , 2011, 2011 IEEE International Conference on Robotics and Automation.

[37]  Jörg Stückler,et al.  The TUM VI Benchmark for Evaluating Visual-Inertial Odometry , 2018, 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[38]  Dongbing Gu,et al.  A review of visual inertial odometry from filtering and optimisation perspectives , 2015, Adv. Robotics.

[39]  Marc Pollefeys,et al.  Semi-direct EKF-based monocular visual-inertial odometry , 2015, 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[40]  Dieter Fox,et al.  DynamicFusion: Reconstruction and tracking of non-rigid scenes in real-time , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[41]  Edwin Olson,et al.  AprilTag: A robust and flexible visual fiducial system , 2011, 2011 IEEE International Conference on Robotics and Automation.

[42]  Daniel Cremers,et al.  Direct Sparse Visual-Inertial Odometry Using Dynamic Marginalization , 2018, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[43]  Timothy Bretl,et al.  Improved Structure from Motion Using Fiducial Marker Matching , 2018, ECCV.

[44]  Yong Zhao,et al.  Map2DFusion: Real-time incremental UAV image mosaicing based on monocular SLAM , 2016, 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[45]  Andreas Zell,et al.  Multi-camera visual SLAM for autonomous navigation of micro aerial vehicles , 2017, Robotics Auton. Syst..

[46]  Salah Sukkarieh,et al.  Visual-Inertial-Aided Navigation for High-Dynamic Motion in Built Environments Without Initial Conditions , 2012, IEEE Transactions on Robotics.

[47]  Daniel Cremers,et al.  LSD-SLAM: Large-Scale Direct Monocular SLAM , 2014, ECCV.

[48]  Daniel Cremers,et al.  Direct Sparse Odometry , 2016, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[49]  Olivier Stasse,et al.  MonoSLAM: Real-Time Single Camera SLAM , 2007, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[50]  Stergios I. Roumeliotis,et al.  A Multi-State Constraint Kalman Filter for Vision-aided Inertial Navigation , 2007, Proceedings 2007 IEEE International Conference on Robotics and Automation.

[51]  Anastasios I. Mourikis,et al.  High-precision, consistent EKF-based visual-inertial odometry , 2013, Int. J. Robotics Res..

[52]  Frank Dellaert,et al.  On-Manifold Preintegration for Real-Time Visual--Inertial Odometry , 2015, IEEE Transactions on Robotics.

[53]  Rafael Muñoz-Salinas,et al.  Mapping and Localization from Planar Markers , 2016, Pattern Recognit..

[54]  Hujun Bao,et al.  Robust Keyframe-based Monocular SLAM for Augmented Reality , 2016, 2016 IEEE International Symposium on Mixed and Augmented Reality (ISMAR).

[55]  Davide Scaramuzza,et al.  SVO: Fast semi-direct monocular visual odometry , 2014, 2014 IEEE International Conference on Robotics and Automation (ICRA).