Cooperative visual-inertial sensor fusion: fundamental equations and state determination in closed-form

This paper investigates the visual and inertial sensor fusion problem in the cooperative case and provides new theoretical and basic results. Specifically, the case of two agents is investigated. Each agent is equipped with inertial sensors (accelerometer and gyroscope) and with a monocular camera. By using the monocular camera, each agent can observe the other agent. No additional camera observations (e.g., of external point features in the environment) are considered. First, the entire observable state is analytically derived. This state contains the relative position between the two agents (which includes the absolute scale), the relative velocity, the three Euler angles that express the rotation between the two local frames and all the accelerometer and gyroscope biases. Then, the basic equations that describe this system are analytically obtained. The last part of the paper describes the use of these equations to obtain a closed-form solution that provides the observable state in terms of the visual and inertial measurements provided in a short time interval. This last contribution is the extension of the results presented in Kaiser et al. (IEEE Robot Autom Lett 2(1):18–25, 2017), Martinelli (IEEE Trans Robot 28(1):44–60, 2012; Int J Comput Vis 106(2):138–152, 2014) to the cooperative case. The impact of the presence of the bias on the performance of this closed-form solution is also investigated and a simple and effective method to obtain the gyroscope bias is proposed. Extensive simulations clearly show that the proposed method is successful. It is worth noting that it is possible to automatically retrieve the absolute scale and simultaneously calibrate the gyroscopes not only without any prior knowledge (as in Kaiser et al. IEEE Robot Autom Lett 2(1):18–25, 2017), but also without external point features in the environment.

[1]  Sonia Martínez,et al.  Cooperative Localization for Mobile Agents: A Recursive Decentralized Algorithm Based on Kalman-Filter Decoupling , 2015, IEEE Control Systems.

[2]  Antonio Franchi,et al.  Bearing rigidity theory in SE(3) , 2016, 2016 IEEE 55th Conference on Decision and Control (CDC).

[3]  Ehud Rivlin,et al.  Graph-based distributed cooperative navigation for a general multi-robot measurement model , 2012, Int. J. Robotics Res..

[4]  Frank Dellaert,et al.  Rigid components identification and rigidity control in bearing-only localization using the graph cycle basis , 2015, 2015 American Control Conference (ACC).

[5]  Stergios I. Roumeliotis,et al.  On the complexity and consistency of UKF-based SLAM , 2009, 2009 IEEE International Conference on Robotics and Automation.

[6]  Michael Bosse,et al.  Keyframe-based visual–inertial odometry using nonlinear optimization , 2015, Int. J. Robotics Res..

[7]  Fawzi Nashashibi,et al.  Cooperative Multi-Vehicle Localization Using Split Covariance Intersection Filter , 2013, IEEE Intelligent Transportation Systems Magazine.

[8]  Agostino Martinelli,et al.  Cooperative visual-inertial sensor fusion: Fundamental equations , 2017, 2017 International Symposium on Multi-Robot and Multi-Agent Systems (MRS).

[9]  Stergios I. Roumeliotis,et al.  An observability-constrained sliding window filter for SLAM , 2011, 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[10]  A. Krener,et al.  Nonlinear controllability and observability , 1977 .

[11]  Stergios I. Roumeliotis,et al.  A Multi-State Constraint Kalman Filter for Vision-aided Inertial Navigation , 2007, Proceedings 2007 IEEE International Conference on Robotics and Automation.

[12]  Anastasios I. Mourikis,et al.  High-precision, consistent EKF-based visual-inertial odometry , 2013, Int. J. Robotics Res..

[13]  Christiaan J. J. Paredis,et al.  Heterogeneous Teams of Modular Robots for Mapping and Exploration , 2000, Auton. Robots.

[14]  Antonio Franchi,et al.  Bearing-only formation control using an SE(2) rigidity theory , 2015, 2015 54th IEEE Conference on Decision and Control (CDC).

[15]  Frank Dellaert,et al.  Rigid Components Identification and Rigidity Enforcement in Bearing-Only Localization using the Graph Cycle Basis , 2015 .

[16]  Stergios I. Roumeliotis,et al.  Decentralized multi-robot cooperative localization using covariance intersection , 2013, 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[17]  Agostino Martinelli,et al.  Closed-Form Solution of Visual-Inertial Structure from Motion , 2013, International Journal of Computer Vision.

[18]  Wolfram Burgard,et al.  Recursive Decentralized Collaborative Localization for Sparsely Communicating Robots , 2016, Robotics: Science and Systems.

[19]  Hiroshi Ishiguro,et al.  Identifying and localizing robots in a multi-robot system environment , 1999, Proceedings 1999 IEEE/RSJ International Conference on Intelligent Robots and Systems. Human and Environment Friendly Robots with High Intelligence and Emotional Quotients (Cat. No.99CH36289).

[20]  Gaurav S. Sukhatme,et al.  Localization for mobile robot teams using maximum likelihood estimation , 2002, IEEE/RSJ International Conference on Intelligent Robots and Systems.

[21]  Davide Scaramuzza,et al.  SVO: Fast semi-direct monocular visual odometry , 2014, 2014 IEEE International Conference on Robotics and Automation (ICRA).

[22]  Dimitrios G. Kottas,et al.  Consistency Analysis and Improvement of Vision-aided Inertial Navigation , 2014, IEEE Transactions on Robotics.

[23]  Salah Sukkarieh,et al.  Visual-Inertial-Aided Navigation for High-Dynamic Motion in Built Environments Without Initial Conditions , 2012, IEEE Transactions on Robotics.

[24]  Gregory Dudek,et al.  Multi-robot cooperative localization: a study of trade-offs between efficiency and accuracy , 2002, IEEE/RSJ International Conference on Intelligent Robots and Systems.

[25]  Roland Siegwart,et al.  Collaborative stereo , 2011, 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[26]  Agostino Martinelli,et al.  Vision and IMU Data Fusion: Closed-Form Solutions for Attitude, Speed, Absolute Scale, and Bias Determination , 2012, IEEE Transactions on Robotics.

[27]  Stergios I. Roumeliotis,et al.  A dual-layer estimator architecture for long-term localization , 2008, 2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops.

[28]  Camillo J. Taylor,et al.  A bounded uncertainty approach to multi-robot localization , 2003, Proceedings 2003 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2003) (Cat. No.03CH37453).

[29]  Nahum Shimkin,et al.  Nonlinear Control Systems , 2008 .

[30]  Joel A. Hesch,et al.  Large-scale cooperative 3D visual-inertial mapping in a Manhattan world , 2016, 2016 IEEE International Conference on Robotics and Automation (ICRA).

[31]  Shaojie Shen,et al.  Spline-Based Initialization of Monocular Visual–Inertial State Estimators at High Altitude , 2017, IEEE Robotics and Automation Letters.

[32]  Stergios I. Roumeliotis,et al.  Robot-to-Robot Relative Pose Estimation From Range Measurements , 2008, IEEE Transactions on Robotics.

[33]  Jiri Matas,et al.  Inertial-based scale estimation for structure from motion on mobile devices , 2017, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[34]  Stergios I. Roumeliotis,et al.  Distributed multirobot localization , 2002, IEEE Trans. Robotics Autom..

[35]  Ioannis M. Rekleitis,et al.  I see you, you see me: Cooperative localization through bearing-only mutually observing robots , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[36]  Markus Vincze,et al.  Fast Ego-motion Estimation with Multi-rate Fusion of Inertial and Vision , 2007, Int. J. Robotics Res..

[37]  Stergios I. Roumeliotis,et al.  Cooperative multi-robot localization under communication constraints , 2009, 2009 IEEE International Conference on Robotics and Automation.

[38]  Keith Yu Kit Leung,et al.  Decentralized Localization of Sparsely-Communicating Robot Networks: A Centralized-Equivalent Approach , 2010, IEEE Transactions on Robotics.

[39]  Frank Dellaert,et al.  Information fusion in navigation systems via factor graph based incremental smoothing , 2013, Robotics Auton. Syst..

[40]  John J. Leonard,et al.  Towards consistent visual-inertial navigation , 2014, 2014 IEEE International Conference on Robotics and Automation (ICRA).

[41]  Agostino Martinelli Closed-form solution to cooperative visual-inertial structure from motion , 2018, ArXiv.

[42]  Frank Dellaert,et al.  IMU Preintegration on Manifold for Efficient Visual-Inertial Maximum-a-Posteriori Estimation , 2015, Robotics: Science and Systems.

[43]  Flavio Fontana,et al.  Simultaneous State Initialization and Gyroscope Bias Calibration in Visual Inertial Aided Navigation , 2017, IEEE Robotics and Automation Letters.

[44]  Stefano Soatto,et al.  Visual-inertial navigation, mapping and localization: A scalable real-time causal approach , 2011, Int. J. Robotics Res..

[45]  Agostino Martinelli,et al.  State Estimation Based on the Concept of Continuous Symmetry and Observability Analysis: The Case of Calibration , 2011, IEEE Transactions on Robotics.

[46]  Luke Fletcher,et al.  Multiple relative pose graphs for robust cooperative mapping , 2010, 2010 IEEE International Conference on Robotics and Automation.