Depth data fusion for simultaneous localization and mapping — RGB-DD SLAM

This paper presents an approach to data fusion from multiple depth sensors with different principles of range measurements. This concept is motivated by the observation that depth sensors exploiting different range measurement techniques have also distinct characteristics of the uncertainty and artifacts in the obtained depth images. Thus, fusing the information from two or more measurement channels allows us to mutually compensate for some of the unwanted effects. The target application for our combined sensor is Simultaneous Localization and Mapping (SLAM). We demonstrated that fusing depth data from two sources in the convex optimization framework yields better results in feature-based 3-D SLAM, than the use of individual sensors for this task. The experimental part is based on data registered with a calibrated rig comprising ASUS Xtion Pro Live and MESA SwissRanger SR-4000 sensors, and ground truth trajectories obtained from a motion capture system. The results of sensor trajectory estimation are demonstrated in terms of the ATE and RPE metrics, widely adopted by the SLAM community.

[1]  Wolfram Burgard,et al.  3-D Mapping With an RGB-D Camera , 2014, IEEE Transactions on Robotics.

[2]  Andreas Zell,et al.  Using depth in visual simultaneous localisation and mapping , 2012, 2012 IEEE International Conference on Robotics and Automation.

[3]  Horst Bischof,et al.  Variational Depth Superresolution Using Example-Based Edge Representations , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).

[4]  Luigi di Stefano,et al.  SlamDunk: Affordable Real-Time RGB-D SLAM , 2014, ECCV Workshops.

[5]  David A. Forsyth,et al.  Sparse depth super resolution , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[6]  Jean-Yves Bouguet,et al.  Camera calibration toolbox for matlab , 2001 .

[7]  Lars Petersson,et al.  Cutting Edge: Soft Correspondences in Multimodal Scene Parsing , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).

[8]  Wolfram Burgard,et al.  G2o: A general framework for graph optimization , 2011, 2011 IEEE International Conference on Robotics and Automation.

[9]  Michal R. Nowicki,et al.  Combining photometric and depth data for lightweight and robust visual odometry , 2013, 2013 European Conference on Mobile Robots.

[10]  John J. Leonard,et al.  Robust real-time visual odometry for dense RGB-D mapping , 2013, 2013 IEEE International Conference on Robotics and Automation.

[11]  Michal R. Nowicki,et al.  Improving accuracy of feature-based RGB-D SLAM by modeling spatial uncertainty of point features , 2016, 2016 IEEE International Conference on Robotics and Automation (ICRA).

[12]  Otmar Loffeld,et al.  Depth Camera Technology Comparison and Performance Evaluation , 2012, ICPRAM.

[13]  Tomás Svoboda,et al.  Robust Data Fusion of Multimodal Sensory Information for Mobile Robots , 2015, J. Field Robotics.

[14]  Dieter Fox,et al.  RGB-D mapping: Using Kinect-style depth cameras for dense 3D modeling of indoor environments , 2012, Int. J. Robotics Res..

[15]  Horst Bischof,et al.  Multi-modality depth map fusion using primal-dual optimization , 2013, IEEE International Conference on Computational Photography (ICCP).

[16]  Daniel Cremers,et al.  Robust odometry estimation for RGB-D cameras , 2013, 2013 IEEE International Conference on Robotics and Automation.

[17]  Sander Oude Elberink,et al.  Accuracy and Resolution of Kinect Depth Data for Indoor Mapping Applications , 2012, Sensors.

[18]  Michal R. Nowicki,et al.  Accurate Map-Based RGB-D SLAM for Mobile Robots , 2015, ROBOT.

[19]  Patrick Pérez,et al.  Incremental dense multi-modal 3D scene reconstruction , 2015, 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[20]  Antonin Chambolle,et al.  A First-Order Primal-Dual Algorithm for Convex Problems with Applications to Imaging , 2011, Journal of Mathematical Imaging and Vision.

[21]  Kurt Konolige,et al.  Double window optimisation for constant time visual SLAM , 2011, 2011 International Conference on Computer Vision.

[22]  Sebastian Thrun,et al.  An Application of Markov Random Fields to Range Sensing , 2005, NIPS.

[23]  Hannes Sommer,et al.  Fusion of optical flow and inertial measurements for robust egomotion estimation , 2014, 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[24]  Piotr Skrzypczynski,et al.  Estimating terrain elevation maps from sparse and uncertain multi-sensor data , 2012, 2012 IEEE International Conference on Robotics and Biomimetics (ROBIO).

[25]  Radu Grosu,et al.  Generic sensor fusion package for ROS , 2015, 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).