Resource-aware Online Parameter Adaptation for Computationally -constrained Visual-Inertial Navigation Systems

In this paper, a computational resources-aware parameter adaptation method for visual-inertial navigation systems is proposed with the goal of enabling the improved deployment of such algorithms on computationally constrained systems. Such a capacity can prove critical when employed on ultra-lightweight systems or alongside mission critical computationally expensive processes. To achieve this objective, the algorithm proposes selected changes in the vision frontend and optimization back-end of visual-inertial odometry algorithms, both prior to execution and in real-time based on an online profiling of available resources. The method also utilizes information from the motion dynamics experienced by the system to manipulate parameters online. The general policy is demonstrated on three established algorithms, namely S-MSCKF, VINS-Mono and OKVIS and has been verified experimentally on the EuRoC dataset. The proposed approach achieved comparable performance at a fraction of the original computational cost.

[1]  Stergios I. Roumeliotis,et al.  A Multi-State Constraint Kalman Filter for Vision-aided Inertial Navigation , 2007, Proceedings 2007 IEEE International Conference on Robotics and Automation.

[2]  Christos Papachristos,et al.  Robust Thermal-Inertial Localization for Aerial Robots: A Case for Direct Methods , 2019, 2019 International Conference on Unmanned Aircraft Systems (ICUAS).

[3]  Peng Wang,et al.  TP-TIO: A Robust Thermal-Inertial Odometry with Deep ThermalPoint , 2020, 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[4]  John J. Leonard,et al.  Consistent sparsification for graph optimization , 2013, 2013 European Conference on Mobile Robots.

[5]  Michael Bosse,et al.  Keyframe-based visual–inertial odometry using nonlinear optimization , 2015, Int. J. Robotics Res..

[6]  Takeo Kanade,et al.  An Iterative Image Registration Technique with an Application to Stereo Vision , 1981, IJCAI.

[7]  Davide Scaramuzza,et al.  A Tutorial on Quantitative Trajectory Evaluation for Visual(-Inertial) Odometry , 2018, 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[8]  Kurt Konolige,et al.  FrameSLAM: From Bundle Adjustment to Real-Time Visual Mapping , 2008, IEEE Transactions on Robotics.

[9]  Lina María Paz,et al.  Divide and Conquer: EKF SLAM in O(n) , 2008, IEEE Trans. Robotics.

[10]  Javier Civera,et al.  Information-Driven Direct RGB-D Odometry , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[11]  John J. Leonard,et al.  Past, Present, and Future of Simultaneous Localization and Mapping: Toward the Robust-Perception Age , 2016, IEEE Transactions on Robotics.

[12]  Woosik Lee,et al.  OpenVINS: A Research Platform for Visual-Inertial Estimation , 2020, 2020 IEEE International Conference on Robotics and Automation (ICRA).

[13]  G. Klein,et al.  Parallel Tracking and Mapping for Small AR Workspaces , 2007, 2007 6th IEEE and ACM International Symposium on Mixed and Augmented Reality.

[14]  Wolfram Burgard,et al.  Which landmark is useful? Learning selection policies for navigation in unknown environments , 2009, 2009 IEEE International Conference on Robotics and Automation.

[15]  Roland Siegwart,et al.  SFly: Swarm of micro flying robots , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[16]  Roland Siegwart,et al.  The EuRoC micro aerial vehicle datasets , 2016, Int. J. Robotics Res..

[17]  Shaojie Shen,et al.  VINS-Mono: A Robust and Versatile Monocular Visual-Inertial State Estimator , 2017, IEEE Transactions on Robotics.

[18]  Vijay Kumar,et al.  Robust Stereo Visual Inertial Odometry for Fast Autonomous Flight , 2017, IEEE Robotics and Automation Letters.

[19]  Frank Dellaert,et al.  IMU Preintegration on Manifold for Efficient Visual-Inertial Maximum-a-Posteriori Estimation , 2015, Robotics: Science and Systems.

[20]  Dimitrios G. Kottas,et al.  A Resource-aware Vision-aided Inertial Navigation System for Wearable and Portable Computers , 2014 .

[21]  W MurrayDavid,et al.  Simultaneous Localization and Map-Building Using Active Vision , 2002 .

[22]  Carlo Tomasi,et al.  Good features to track , 1994, 1994 Proceedings of IEEE Conference on Computer Vision and Pattern Recognition.

[23]  Luca Carlone,et al.  Visual-Inertial Odometry on Chip: An Algorithm-and-Hardware Co-design Approach , 2017, Robotics: Science and Systems.

[24]  Roland Siegwart,et al.  BRISK: Binary Robust invariant scalable keypoints , 2011, 2011 International Conference on Computer Vision.

[25]  Richard Elvira,et al.  ORB-SLAM3: An Accurate Open-Source Library for Visual, Visual–Inertial, and Multimap SLAM , 2021, IEEE Transactions on Robotics.

[26]  Anastasios I. Mourikis,et al.  Vision-aided inertial navigation for resource-constrained systems , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[27]  Christos Papachristos,et al.  Keyframe-based Direct Thermal–Inertial Odometry , 2019, 2019 International Conference on Robotics and Automation (ICRA).

[28]  Mark Hedley,et al.  Fast corner detection , 1998, Image Vis. Comput..

[29]  Frank Dellaert,et al.  Eliminating conditionally independent sets in factor graphs: A unifying perspective based on smart factors , 2014, 2014 IEEE International Conference on Robotics and Automation (ICRA).