Balancing the Budget: Feature Selection and Tracking for Multi-Camera Visual-Inertial Odometry

We present a multi-camera visual-inertial odometry system based on factor graph optimization which estimates motion by using all cameras simultaneously while retaining a fixed overall feature budget. We focus on motion tracking in challenging environments, such as narrow corridors, dark spaces with aggressive motions, and abrupt lighting changes. These scenarios cause traditional monocular or stereo odometry to fail. While tracking motion with extra cameras should theoretically prevent failures, it leads to additional complexity and computational burden. To overcome these challenges, we introduce two novel methods to improve multi-camera feature tracking. First, instead of tracking features separately in each camera, we track features continuously as they move from one camera to another. This increases accuracy and achieves a more compact factor graph representation. Second, we select a fixed budget of tracked features across the cameras to reduce back-end optimization time. We have found that using a smaller set of informative features can maintain the same tracking accuracy. Our proposed method was extensively tested using a hardware-synchronized device consisting of an IMU and four cameras (a front stereo pair and two lateral) in scenarios including: an underground mine, large open spaces, and building interiors with narrow stairs and corridors. Compared to stereo-only state-of-the-art visual-inertial odometry methods, our approach reduces the drift rate, relative pose error, by up to 80% in translation and 39% in rotation.

[1]  Woosik Lee,et al.  OpenVINS: A Research Platform for Visual-Inertial Estimation , 2020, 2020 IEEE International Conference on Robotics and Automation (ICRA).

[2]  Yipu Zhao,et al.  Good Feature Selection for Least Squares Pose Optimization in VO/VSLAM , 2018, 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[3]  Milad Ramezani,et al.  The Newer College Dataset: Handheld LiDAR, Inertial and Vision with Ground Truth , 2020, ArXiv.

[4]  Bernhard P. Wrobel,et al.  Multiple View Geometry in Computer Vision , 2001 .

[5]  Yipu Zhao,et al.  Good Feature Matching: Toward Accurate, Robust VO/VSLAM With Low Latency , 2020, IEEE Transactions on Robotics.

[6]  Luca Carlone,et al.  Attention and anticipation in fast visual-inertial navigation , 2016, 2017 IEEE International Conference on Robotics and Automation (ICRA).

[7]  Maurice Fallon,et al.  Unified Multi-Modal Landmark Tracking for Tightly Coupled Lidar-Visual-Inertial Odometry , 2020, IEEE Robotics and Automation Letters.

[8]  Torsten Sattler,et al.  Towards Robust Visual Odometry with a Multi-Camera System , 2018, 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[9]  Jongwoo Lim,et al.  ROVINS: Robust Omnidirectional Visual Inertial Navigation System , 2020, IEEE Robotics and Automation Letters.

[10]  Davide Scaramuzza,et al.  A Benchmark Comparison of Monocular Visual-Inertial Odometry Algorithms for Flying Robots , 2018, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[11]  Peter Fankhauser,et al.  ANYmal - a highly mobile and dynamic quadrupedal robot , 2016, 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[12]  Kevin Eckenhoff,et al.  MIMC-VINS: A Versatile and Resilient Multi-IMU Multi-Camera Visual-Inertial Navigation System , 2020, IEEE Transactions on Robotics.

[13]  Roland Siegwart,et al.  CERBERUS: Autonomous Legged and Aerial Robotic Exploration in the Tunnel and Urban Circuits of the DARPA Subterranean Challenge , 2022, Field Robotics.

[14]  Michael Gassner,et al.  SVO: Semidirect Visual Odometry for Monocular and Multicamera Systems , 2017, IEEE Transactions on Robotics.

[15]  Roland Siegwart,et al.  Extending kalibr: Calibrating the extrinsics of multiple IMUs and of individual axes , 2016, 2016 IEEE International Conference on Robotics and Automation (ICRA).

[16]  Juan D. Tardós,et al.  ORB-SLAM2: An Open-Source SLAM System for Monocular, Stereo, and RGB-D Cameras , 2016, IEEE Transactions on Robotics.

[17]  Teodor Tomic,et al.  Robust Visual-Inertial State Estimation with Multiple Odometries and Efficient Mapping on an MAV with Ultra-Wide FOV Stereo Vision , 2018, 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[18]  D. Scaramuzza,et al.  Redesigning SLAM for Arbitrary Multi-Camera Systems , 2020, 2020 IEEE International Conference on Robotics and Automation (ICRA).

[19]  Joshua G. Mangelson,et al.  A Robust Multi-Stereo Visual-Inertial Odometry Pipeline , 2020, 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[20]  Michael Bosse,et al.  Keyframe-based visual–inertial odometry using nonlinear optimization , 2015, Int. J. Robotics Res..

[21]  Frank Dellaert,et al.  iSAM2: Incremental smoothing and mapping with fluid relinearization and incremental variable reordering , 2011, 2011 IEEE International Conference on Robotics and Automation.

[22]  Frank Dellaert,et al.  On-Manifold Preintegration for Real-Time Visual--Inertial Odometry , 2015, IEEE Transactions on Robotics.

[23]  Richard Elvira,et al.  ORB-SLAM3: An Accurate Open-Source Library for Visual, Visual–Inertial, and Multimap SLAM , 2021, IEEE Transactions on Robotics.

[24]  Tim D. Barfoot,et al.  At all Costs: A Comparison of Robust Cost Functions for Camera Correspondence Outliers , 2015, 2015 12th Conference on Computer and Robot Vision.

[25]  Andreas Krause,et al.  Lazier Than Lazy Greedy , 2014, AAAI.

[26]  Qian Huang,et al.  A review of monocular visual odometry , 2019, The Visual Computer.

[27]  Shaojie Shen,et al.  VINS-Mono: A Robust and Versatile Monocular Visual-Inertial State Estimator , 2017, IEEE Transactions on Robotics.

[28]  Frank Dellaert,et al.  Factor Graphs for Robot Perception , 2017, Found. Trends Robotics.

[29]  Andrew Zisserman,et al.  Multiple View Geometry in Computer Vision (2nd ed) , 2003 .

[30]  Maurice Fallon,et al.  VILENS: Visual, Inertial, Lidar, and Leg Odometry for All-Terrain Legged Robots , 2021, ArXiv.