A Robust Indoor/Outdoor Navigation Filter Fusing Data from Vision and Magneto-Inertial Measurement Unit

Visual-inertial Navigation Systems (VINS) are nowadays used for robotic or augmented reality applications. They aim to compute the motion of the robot or the pedestrian in an environment that is unknown and does not have specific localization infrastructure. Because of the low quality of inertial sensors that can be used reasonably for these two applications, state of the art VINS rely heavily on the visual information to correct at high frequency the drift of inertial sensors integration. These methods struggle when environment does not provide usable visual features, such than in low-light of texture-less areas. In the last few years, some work have been focused on using an array of magnetometers to exploit opportunistic stationary magnetic disturbances available indoor in order to deduce a velocity. This led to Magneto-inertial Dead-reckoning (MI-DR) systems that show interesting performance in their nominal conditions, even if they can be defeated when the local magnetic gradient is too low, for example outdoor. We propose in this work to fuse the information from a monocular camera with the MI-DR technique to increase the robustness of both traditional VINS and MI-DR itself. We use an inverse square root filter inspired by the MSCKF algorithm and describe its structure thoroughly in this paper. We show navigation results on a real dataset captured by a sensor fusing a commercial-grade camera with our custom MIMU (Magneto-inertial Measurment Unit) sensor. The fused estimate demonstrates higher robustness compared to pure VINS estimate, specially in areas where vision is non informative. These results could ultimately increase the working domain of mobile augmented reality systems.

[1]  Petros G. Voulgaris,et al.  On optimal ℓ∞ to ℓ∞ filtering , 1995, Autom..

[2]  Stergios I. Roumeliotis,et al.  A Square Root Inverse Filter for Efficient Vision-aided Inertial Navigation on Mobile Devices , 2015, Robotics: Science and Systems.

[3]  Nicolas Petit,et al.  Iterative calibration method for inertial and magnetic sensors , 2009, Proceedings of the 48h IEEE Conference on Decision and Control (CDC) held jointly with 2009 28th Chinese Control Conference.

[4]  Jörg Stückler,et al.  Direct visual-inertial odometry with stereo cameras , 2016, 2016 IEEE International Conference on Robotics and Automation (ICRA).

[5]  Ji Zhang,et al.  Real-time depth enhanced monocular odometry , 2014, 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[6]  Guy Le Besnerais,et al.  Infrastructureless indoor navigation with an hybrid magneto-inertial and depth sensor system , 2016, 2016 International Conference on Indoor Positioning and Indoor Navigation (IPIN).

[7]  Stergios I. Roumeliotis,et al.  A Multi-State Constraint Kalman Filter for Vision-aided Inertial Navigation , 2007, Proceedings 2007 IEEE International Conference on Robotics and Automation.

[8]  Anastasios I. Mourikis,et al.  High-precision, consistent EKF-based visual-inertial odometry , 2013, Int. J. Robotics Res..

[9]  Shaojie Shen,et al.  Monocular Visual–Inertial State Estimation With Online Initialization and Camera–IMU Extrinsic Calibration , 2017, IEEE Transactions on Automation Science and Engineering.

[10]  Stefano Soatto,et al.  Observability, identifiability and sensitivity of vision-aided inertial navigation , 2015, 2015 IEEE International Conference on Robotics and Automation (ICRA).

[11]  Roland Siegwart,et al.  Unified temporal and spatial calibration for multi-sensor systems , 2013, 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[12]  D. Simon Kalman filtering with state constraints: a survey of linear and nonlinear algorithms , 2010 .

[13]  Joel A. Hesch,et al.  A comparative analysis of tightly-coupled monocular, binocular, and stereo VINS , 2017, 2017 IEEE International Conference on Robotics and Automation (ICRA).

[14]  Stergios I. Roumeliotis,et al.  IMU-RGBD camera navigation using point and plane features , 2013, 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[15]  Guy Le Besnerais,et al.  An inverse square root filter for robust indoor/outdoor magneto-visual-inertial odometry , 2017, 2017 International Conference on Indoor Positioning and Indoor Navigation (IPIN).

[16]  Nicolas Petit,et al.  Combining inertial measurements and distributed magnetometry for motion estimation , 2011, Proceedings of the 2011 American Control Conference.

[17]  Kurt Konolige,et al.  Large-Scale Visual Odometry for Rough Terrain , 2007, ISRR.

[18]  Dimitrios G. Kottas,et al.  Camera-IMU-based localization: Observability analysis and consistency improvement , 2014, Int. J. Robotics Res..

[19]  Guy Le Besnerais,et al.  Robust indoor/outdoor navigation through magneto-visual-inertial optimization-based estimation , 2017, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[20]  Anastasios I. Mourikis,et al.  Estimator initialization in vision-aided inertial navigation with unknown camera-IMU calibration , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[21]  Christophe Prieur,et al.  Motion estimation of a rigid body with an EKF using magneto-inertial measurements , 2016, 2016 International Conference on Indoor Positioning and Indoor Navigation (IPIN).

[22]  Frank Dellaert,et al.  On-Manifold Preintegration Theory for Fast and Accurate Visual-Inertial Navigation , 2015, ArXiv.

[23]  Michael Bosse,et al.  Keyframe-based visual–inertial odometry using nonlinear optimization , 2015, Int. J. Robotics Res..

[24]  Christophe Prieur,et al.  Improving magneto-inertial attitude and position estimation by means of a magnetic heading observer , 2017, 2017 International Conference on Indoor Positioning and Indoor Navigation (IPIN).