Moving object detection for visual odometry in a dynamic environment based on occlusion accumulation

Detection of moving objects is an essential capability in dealing with dynamic environments. Most moving object detection algorithms have been designed for color images without depth. For robotic navigation where real-time RGBD data is often readily available, utilization of the depth information would be beneficial for obstacle recognition. Here, we propose a simple moving object detection algorithm that uses RGB-D images. The proposed algorithm does not require estimating a background model. Instead, it uses an occlusion model which enables us to estimate the camera pose on a background confused with moving objects that dominate the scene. The proposed algorithm allows to separate the moving object detection and visual odometry (VO) so that an arbitrary robust VO method can be employed in a dynamic situation with a combination of moving object detection, whereas other VO algorithms for a dynamic environment are inseparable. In this paper, we use dense visual odometry (DVO) as a VO method with a bi-square regression weight. Experimental results show the segmentation accuracy and the performance improvement of DVO in the situations. We validate our algorithm in public datasets and our dataset which also publicly accessible.

[1]  Daniel Cremers,et al.  Fast odometry and scene flow from RGB-D cameras based on geometric clustering , 2017, 2017 IEEE International Conference on Robotics and Automation (ICRA).

[2]  Young-Jun Son,et al.  Effective and Efficient Detection of Moving Targets From a UAV’s Camera , 2018, IEEE Transactions on Intelligent Transportation Systems.

[3]  Hyung Jin Chang,et al.  Detection of Moving Objects with Non-stationary Cameras in 5.8ms: Bringing Motion Detection to Your Mobile Device , 2013, 2013 IEEE Conference on Computer Vision and Pattern Recognition Workshops.

[4]  Krishnan Kutty,et al.  Background Modelling from a Moving Camera , 2015 .

[5]  Christian Kerl Odometry from RGB-D Cameras for Autonomous Quadrocopters , 2012 .

[6]  Yu Liu,et al.  Background Subtraction for Moving Cameras Based on Trajectory-controlled Segmentation and Label Inference , 2015, KSII Trans. Internet Inf. Syst..

[7]  H. Jin Kim,et al.  Real-time rigid motion segmentation using grid-based optical flow , 2017, 2017 IEEE International Conference on Systems, Man, and Cybernetics (SMC).

[8]  Jitendra Malik,et al.  Object Segmentation by Long Term Analysis of Point Trajectories , 2010, ECCV.

[9]  Daniel Cremers,et al.  Robust odometry estimation for RGB-D cameras , 2013, 2013 IEEE International Conference on Robotics and Automation.

[10]  Mehran Yazdi,et al.  New trends on moving object detection in video images captured by a moving camera: A survey , 2018, Comput. Sci. Rev..

[11]  Alper Yilmaz,et al.  Background subtraction for the moving camera: A geometric approach , 2014, Comput. Vis. Image Underst..

[12]  Yuxiang Sun,et al.  Motion removal from moving platforms: An RGB-D data-based motion detection, tracking and segmentation approach , 2015, 2015 IEEE International Conference on Robotics and Biomimetics (ROBIO).

[13]  Takeo Kanade,et al.  Background Subtraction for Freely Moving Cameras , 2009, 2009 IEEE 12th International Conference on Computer Vision.

[14]  Jong-Hwan Kim,et al.  Effective Background Model-Based RGB-D Dense Visual Odometry in a Dynamic Environment , 2016, IEEE Transactions on Robotics.

[15]  Soon Ki Jung,et al.  Moving Object Detection on RGB-D Videos Using Graph Regularized Spatiotemporal RPCA , 2017, ICIAP Workshops.

[16]  Wu-Chih Hu,et al.  Moving object detection and tracking from video captured by moving camera , 2015, J. Vis. Commun. Image Represent..

[17]  John J. Leonard,et al.  Robust real-time visual odometry for dense RGB-D mapping , 2013, 2013 IEEE International Conference on Robotics and Automation.

[18]  Wolfram Burgard,et al.  A benchmark for the evaluation of RGB-D SLAM systems , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.