An inverse square root filter for robust indoor/outdoor magneto-visual-inertial odometry

We explore a new sensor suite to provide a precise and robust navigation information, primarily intended for pedestrian localisation. We use an IMU sensor augmented with an array of magnetometers, called MIMU (for Magneto-Inertial measurement Unit) hereafter, and a single central camera as the vision sensor. The MIMU sensor has been shown in previous work to significantly improve the inertial dead-reckoning for pedestrian, provided we can assume a stationary and non-uniform magnetic environment. Such conditions are encountered in particular in indoor environment, where visual methods can be challenged by textureless or low-light areas. Alternatively, visual methods are useful in outdoor trajectories, where MIMU often fails because of a uniform magnetic field. We build a fusion estimator based on an inverse square-root filter fed with the raw measurement from magnetometers, accelerometers, gyrometers and visual features extracted and tracked in the video. Real-data experiments show the benefits of the fusion : on the one hand, the added information from the MIMU allows to complement the vision/inertial system in case where the vision does not provide useful information during an extended period of time; on the other hand, vision does extend the operational domain of the navigation system compared to a pure MIMU solution, in particular for outdoor sections of the trajectory.

[1]  Nicolas Petit,et al.  Iterative calibration method for inertial and magnetic sensors , 2009, Proceedings of the 48h IEEE Conference on Decision and Control (CDC) held jointly with 2009 28th Chinese Control Conference.

[2]  Petros G. Voulgaris,et al.  On optimal ℓ∞ to ℓ∞ filtering , 1995, Autom..

[3]  Kurt Konolige,et al.  Large-Scale Visual Odometry for Rough Terrain , 2007, ISRR.

[4]  Christophe Prieur,et al.  Motion estimation of a rigid body with an EKF using magneto-inertial measurements , 2016, 2016 International Conference on Indoor Positioning and Indoor Navigation (IPIN).

[5]  Guy Le Besnerais,et al.  Infrastructureless indoor navigation with an hybrid magneto-inertial and depth sensor system , 2016, 2016 International Conference on Indoor Positioning and Indoor Navigation (IPIN).

[6]  Michael Bosse,et al.  Keyframe-based visual–inertial odometry using nonlinear optimization , 2015, Int. J. Robotics Res..

[7]  Frank Dellaert,et al.  On-Manifold Preintegration Theory for Fast and Accurate Visual-Inertial Navigation , 2015, ArXiv.

[8]  Stergios I. Roumeliotis,et al.  A Square Root Inverse Filter for Efficient Vision-aided Inertial Navigation on Mobile Devices , 2015, Robotics: Science and Systems.

[9]  Stergios I. Roumeliotis,et al.  IMU-RGBD camera 3D pose estimation and extrinsic calibration: Observability analysis and consistency improvement , 2013, 2013 IEEE International Conference on Robotics and Automation.

[10]  Nicolas Petit,et al.  Combining inertial measurements and distributed magnetometry for motion estimation , 2011, Proceedings of the 2011 American Control Conference.

[11]  Roland Siegwart,et al.  Unified temporal and spatial calibration for multi-sensor systems , 2013, 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[12]  Stergios I. Roumeliotis,et al.  A Multi-State Constraint Kalman Filter for Vision-aided Inertial Navigation , 2007, Proceedings 2007 IEEE International Conference on Robotics and Automation.

[13]  Anastasios I. Mourikis,et al.  High-precision, consistent EKF-based visual-inertial odometry , 2013, Int. J. Robotics Res..

[14]  Ji Zhang,et al.  Real-time depth enhanced monocular odometry , 2014, 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[15]  Stefano Soatto,et al.  Observability, identifiability and sensitivity of vision-aided inertial navigation , 2015, 2015 IEEE International Conference on Robotics and Automation (ICRA).