Consistent Monocular Ackermann Visual–Inertial Odometry for Intelligent and Connected Vehicle Localization

The observability of the scale direction in visual–inertial odometry (VIO) under degenerate motions of intelligent and connected vehicles can be improved by fusing Ackermann error state measurements. However, the relative kinematic error measurement model assumes that the vehicle velocity is constant between two consecutive camera states, which degrades the positioning accuracy. To address this problem, a consistent monocular Ackermann VIO, termed MAVIO, is proposed to combine the vehicle velocity and yaw angular rate error measurements, taking into account the lever arm effect between the vehicle and inertial measurement unit (IMU) coordinates with a tightly coupled filter-based mechanism. The lever arm effect is firstly introduced to improve the reliability for information exchange between the vehicle and IMU coordinates. Then, the process model and monocular visual measurement model are presented. Subsequently, the vehicle velocity and yaw angular rate error measurements are directly used to refine the estimator after visual observation. To obtain a global position for the vehicle, the raw Global Navigation Satellite System (GNSS) error measurement model, termed MAVIO-GNSS, is introduced to further improve the performance of MAVIO. The observability, consistency and positioning accuracy were comprehensively compared using real-world datasets. The experimental results demonstrated that MAVIO not only improved the observability of the VIO scale direction under the degenerate motions of ground vehicles, but also resolved the inconsistency problem of the relative kinematic error measurement model of the vehicle to further improve the positioning accuracy. Moreover, MAVIO-GNSS further improved the vehicle positioning accuracy under a long-distance driving state. The source code is publicly available for the benefit of the robotics community.

[1]  Fangwu Ma,et al.  ACK-MSCKF: Tightly-Coupled Ackermann Multi-State Constraint Kalman Filter for Autonomous Vehicle Localization , 2019, Sensors.

[2]  Gang Peng,et al.  Robust tightly coupled pose estimation based on monocular vision, inertia, and wheel speed , 2020, ArXiv.

[3]  Stergios I. Roumeliotis,et al.  A Square Root Inverse Filter for Efficient Vision-aided Inertial Navigation on Mobile Devices , 2015, Robotics: Science and Systems.

[4]  Peng Gang,et al.  Robust tightly coupled pose estimation based on monocular vision, inertia, and wheel speed , 2020 .

[5]  Dimitrios G. Kottas,et al.  Consistency Analysis and Improvement of Vision-aided Inertial Navigation , 2014, IEEE Transactions on Robotics.

[6]  Juan D. Tardós,et al.  ORB-SLAM2: An Open-Source SLAM System for Monocular, Stereo, and RGB-D Cameras , 2016, IEEE Transactions on Robotics.

[7]  Jiawei Wang,et al.  Stability Design for the Homogeneous Platoon with Communication Time Delay , 2020 .

[8]  Stergios I. Roumeliotis,et al.  Unobservable Directions of VINS Under Special Motions , 2016 .

[9]  Mingming Zhang,et al.  Vision-Aided Localization For Ground Robots , 2019, 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[10]  Wenxing Fu,et al.  Lightweight hybrid visual-inertial odometry with closed-form zero velocity update , 2020 .

[11]  Guoquan Huang,et al.  Visual-Inertial Navigation: A Concise Review , 2019, 2019 International Conference on Robotics and Automation (ICRA).

[12]  Kevin M. Lynch,et al.  Modern Robotics: Mechanics, Planning, and Control , 2017 .

[13]  Stergios I. Roumeliotis,et al.  VINS on wheels , 2017, 2017 IEEE International Conference on Robotics and Automation (ICRA).

[14]  J.L. Crassidis,et al.  Sigma-point Kalman filtering for integrated GPS and inertial navigation , 2005, IEEE Transactions on Aerospace and Electronic Systems.

[15]  Vijay Kumar,et al.  Robust Stereo Visual Inertial Odometry for Fast Autonomous Flight , 2017, IEEE Robotics and Automation Letters.

[16]  Dimitrios G. Kottas,et al.  Observability-constrained Vision-aided Inertial Navigation , 2012 .

[17]  Davide Scaramuzza,et al.  Visual-Inertial Odometry of Aerial Robots , 2019, ArXiv.

[18]  E. Grafarend The Optimal Universal Transverse Mercator Projection , 1995 .

[19]  Hujun Bao,et al.  ICE-BA: Incremental, Consistent and Efficient Bundle Adjustment for Visual-Inertial SLAM , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[20]  Zhanyi Hu,et al.  Bidirectional Trajectory Computation for Odometer-Aided Visual-Inertial SLAM , 2020, ArXiv.

[21]  Carlos Campos,et al.  ORB-SLAM3: An Accurate Open-Source Library for Visual, Visual–Inertial, and Multimap SLAM , 2020, IEEE Transactions on Robotics.

[22]  Kevin Eckenhoff,et al.  Intermittent GPS-aided VIO: Online Initialization and Calibration , 2020, 2020 IEEE International Conference on Robotics and Automation (ICRA).

[23]  Y. Oshman,et al.  Averaging Quaternions , 2007 .

[24]  Kanwar Bharat Singh,et al.  Literature review and fundamental approaches for vehicle and tire state estimation* , 2018, Vehicle System Dynamics.

[25]  Shaojie Shen,et al.  A General Optimization-based Framework for Local Odometry Estimation with Multiple Sensors , 2019, ArXiv.

[26]  Daniel Cremers,et al.  Direct Sparse Visual-Inertial Odometry Using Dynamic Marginalization , 2018, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[27]  Roland Siegwart,et al.  Robust visual inertial odometry using a direct EKF-based approach , 2015, IROS 2015.

[28]  Ji Zhao,et al.  PL-VIO: Tightly-Coupled Monocular Visual–Inertial Odometry Using Point and Line Features , 2018, Sensors.

[29]  Yun-Hui Liu,et al.  SE(2)-Constrained Visual Inertial Fusion for Ground Vehicles , 2018, IEEE Sensors Journal.

[30]  Shaojie Shen,et al.  VINS-Mono: A Robust and Versatile Monocular Visual-Inertial State Estimator , 2017, IEEE Transactions on Robotics.

[31]  Michael Bosse,et al.  Keyframe-based visual–inertial odometry using nonlinear optimization , 2015, Int. J. Robotics Res..

[32]  Wei Gao,et al.  Visual-Inertial Odometry Tightly Coupled with Wheel Encoder Adopting Robust Initialization and Online Extrinsic Calibration , 2019, 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[33]  J. M. M. Montiel,et al.  ORB-SLAM: A Versatile and Accurate Monocular SLAM System , 2015, IEEE Transactions on Robotics.

[34]  Mingyang Li,et al.  Visual-Inertial Localization for Skid-Steering Robots with Kinematic Constraints , 2019, ISRR.

[35]  Tianmiao Wang,et al.  Tightly-coupled Data Fusion of VINS and Odometer Based on Wheel Slip Estimation , 2018, 2018 IEEE International Conference on Robotics and Biomimetics (ROBIO).

[36]  Keyvan Ansari,et al.  Cooperative Position Prediction: Beyond Vehicle-to-Vehicle Relative Positioning , 2020, IEEE Transactions on Intelligent Transportation Systems.

[37]  Yong Liu,et al.  Robust and Efficient Vehicles Motion Estimation with Low-Cost Multi-Camera and Odometer-Gyroscope , 2019, 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[38]  Stergios I. Roumeliotis,et al.  A Multi-State Constraint Kalman Filter for Vision-aided Inertial Navigation , 2007, Proceedings 2007 IEEE International Conference on Robotics and Automation.

[39]  Cezary Specht,et al.  Road Tests of the Positioning Accuracy of INS/GNSS Systems Based on MEMS Technology for Navigating Railway Vehicles , 2020 .

[40]  Guoquan Huang,et al.  Robocentric visual–inertial odometry , 2018, 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[41]  Stefan Leutenegger,et al.  KO-Fusion: Dense Visual SLAM with Tightly-Coupled Kinematic and Odometric Tracking , 2019, 2019 International Conference on Robotics and Automation (ICRA).

[42]  Roland Siegwart,et al.  A robust and modular multi-sensor fusion approach applied to MAV navigation , 2013, 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[43]  Rong Kang,et al.  VINS-Vehicle: A Tightly-Coupled Vehicle Dynamics Extension to Visual-Inertial State Estimator , 2019, 2019 IEEE Intelligent Transportation Systems Conference (ITSC).

[44]  Shi-Sheng Huang,et al.  Tightly-Coupled Monocular Visual-Odometric SLAM Using Wheels and a MEMS Gyroscope , 2018, IEEE Access.

[45]  Guoquan Huang,et al.  Degenerate Motion Analysis for Aided INS With Online Spatial and Temporal Sensor Calibration , 2019, IEEE Robotics and Automation Letters.

[46]  Kun Jiang,et al.  A Unified Multiple-Target Positioning Framework for Intelligent Connected Vehicles , 2019, Sensors.

[47]  Shaojie Shen,et al.  A General Optimization-based Framework for Global Pose Estimation with Multiple Sensors , 2019, ArXiv.

[48]  Kourosh Khoshelham,et al.  Vehicle Positioning in GNSS-Deprived Urban Areas by Stereo Visual-Inertial Odometry , 2018, IEEE Transactions on Intelligent Vehicles.

[49]  Kun Jiang,et al.  Intelligent and connected vehicles: Current status and future perspectives , 2018, Science China Technological Sciences.

[50]  Woosik Lee,et al.  Visual-Inertial-Wheel Odometry with Online Calibration , 2020, 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[51]  S. Umeyama,et al.  Least-Squares Estimation of Transformation Parameters Between Two Point Patterns , 1991, IEEE Trans. Pattern Anal. Mach. Intell..

[52]  Rong Xiong,et al.  Gyro-aided camera-odometer online calibration and localization , 2017, 2017 American Control Conference (ACC).

[53]  Mingming Jiang,et al.  A New Filtering and Smoothing Algorithm for Railway Track Surveying Based on Landmark and IMU/Odometer , 2017, Sensors.

[54]  Jörg Stückler,et al.  Visual-Inertial Mapping With Non-Linear Factor Recovery , 2019, IEEE Robotics and Automation Letters.

[55]  Chenxu Zhao,et al.  Monocular Visual-Inertial Odometry with an Unbiased Linear System Model and Robust Feature Tracking Front-End , 2019, Sensors.

[56]  N. Trawny,et al.  Indirect Kalman Filter for 3 D Attitude Estimation , 2005 .

[57]  Guoquan Huang,et al.  A Linear-Complexity EKF for Visual-Inertial Navigation with Loop Closures , 2019, 2019 International Conference on Robotics and Automation (ICRA).

[58]  Juan D. Tardós,et al.  Visual-Inertial Monocular SLAM With Map Reuse , 2016, IEEE Robotics and Automation Letters.

[59]  Davide Scaramuzza,et al.  A Tutorial on Quantitative Trajectory Evaluation for Visual(-Inertial) Odometry , 2018, 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).