I-MOVE: Independent Moving Objects for Velocity Estimation

We introduce I-MOVE, the first publicly available RGBD/stereo dataset for estimating velocities of independently moving objects. Velocity estimation given RGB-D data is an unsolved problem. The I-MOVE dataset provides an opportunity for generalizable velocity estimation models to be created and have their performance be accurately and fairly measured. The dataset features various outdoor and indoor scenes of single and multiple moving objects. Compared to other datasets, I-MOVE is unique because the RGB-D data and speed for each object are supplied for a variety of different settings/environments, objects, and motions. The dataset includes training and test sequences captured from four different depth camera views and three 4Kstereo setups. The data are also time-synchronized with three Doppler radars to provide the magnitude of velocity ground truth. The I-MOVE dataset includes complex scenes from moving pedestrians via walking and biking to multiple rolling objects, all captured with the seven cameras, providing over 500 Depth/Stereo videos. To access the dataset please visit www.vast.uccs.edu/imove

[1]  Pierre Pettersson,et al.  Estimation of Vehicle Lateral Velocity , 2008 .

[2]  Mandyam V. Srinivasan,et al.  Vision-only egomotion estimation in 6DOF using a sky compass , 2018, Robotica.

[3]  Chao Deng,et al.  Multiobject Tracking in Videos Based on LSTM and Deep Reinforcement Learning , 2018, Complex..

[4]  Wei Liu,et al.  Automated Vehicle Attitude and Lateral Velocity Estimation Using a 6-D IMU Aided by Vehicle Dynamics , 2018, 2018 IEEE Intelligent Vehicles Symposium (IV).

[5]  Petros Daras,et al.  A Computerized System for Real-Time Exercise Performance Monitoring and e-Coaching Using Motion Capture Data , 2018 .

[6]  Nuno Vasconcelos,et al.  Person-following UAVs , 2016, 2016 IEEE Winter Conference on Applications of Computer Vision (WACV).

[7]  Jiri Matas,et al.  Discriminative Correlation Filter with Channel and Spatial Reliability , 2017, CVPR.

[8]  Michael S. Ryoo,et al.  Representation Flow for Action Recognition , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[9]  Jürgen Schmidhuber,et al.  Kalman filters improve LSTM network performance in problems unsolvable by traditional recurrent nets , 2003, Neural Networks.

[10]  Yi Yang,et al.  Adversarial Complementary Learning for Weakly Supervised Object Localization , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[11]  Michael J. Black,et al.  HumanEva: Synchronized Video and Motion Capture Dataset and Baseline Algorithm for Evaluation of Articulated Human Motion , 2010, International Journal of Computer Vision.

[12]  Zhiqiang Gao,et al.  An adaptive nonlinear filter approach to the vehicle velocity estimation for ABS , 2000, Proceedings of the 2000. IEEE International Conference on Control Applications. Conference Proceedings (Cat. No.00CH37162).

[13]  Silvio Savarese,et al.  SURREAL: Open-Source Reinforcement Learning Framework and Robot Manipulation Benchmark , 2018, CoRL.

[14]  Tarak Gandhi,et al.  Pedestrian collision avoidance systems: a survey of computer vision based recent studies , 2006, 2006 IEEE Intelligent Transportation Systems Conference.

[15]  Nassir Navab,et al.  Long Short-Term Memory Kalman Filters: Recurrent Neural Estimators for Pose Regularization , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[16]  Niklas Bergström,et al.  Velocity Estimation for UAVs by Using High-Speed Vision , 2018, J. Robotics Mechatronics.

[17]  Dimitrios Zorbas,et al.  Energy Efficient Mobile Target Tracking Using Flying Drones , 2013, ANT/SEIT.

[18]  Trevor Darrell,et al.  Accurate Visual Localization for Automotive Applications , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).

[19]  James M. Rehg,et al.  Multi-object Tracking with Neural Gating Using Bilinear LSTM , 2018, ECCV.

[20]  Davide Anguita,et al.  A Public Domain Dataset for Human Activity Recognition using Smartphones , 2013, ESANN.

[21]  Hikaru Inooka,et al.  Real-time algorithms for estimating jerk signals from noisy acceleration data , 2003 .

[22]  Stefan Roth,et al.  MOT16: A Benchmark for Multi-Object Tracking , 2016, ArXiv.

[23]  Vijay Kumar,et al.  The Multivehicle Stereo Event Camera Dataset: An Event Camera Dataset for 3D Perception , 2018, IEEE Robotics and Automation Letters.

[24]  Petros Daras,et al.  Three-dimensional monitoring of weightlifting for computer assisted training , 2013, VRIC.

[25]  Alejandro Hernández Cordero,et al.  Extending the OpenAI Gym for robotics: a toolkit for reinforcement learning using ROS and Gazebo , 2016, ArXiv.

[26]  Marc Pollefeys,et al.  Motion Estimation for Self-Driving Cars with a Generalized Camera , 2013, 2013 IEEE Conference on Computer Vision and Pattern Recognition.

[27]  Qiang Xu,et al.  nuScenes: A Multimodal Dataset for Autonomous Driving , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[28]  Kris M. Kitani,et al.  Predicting wide receiver trajectories in American football , 2016, 2016 IEEE Winter Conference on Applications of Computer Vision (WACV).

[29]  Kwanghoon Sohn,et al.  Deep Monocular Depth Estimation via Integration of Global and Local Predictions , 2018, IEEE Transactions on Image Processing.

[30]  Silvio Savarese,et al.  Tracking the Untrackable: Learning to Track Multiple Cues with Long-Term Dependencies , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[31]  Osama Masoud,et al.  A vision-based approach to collision prediction at traffic intersections , 2005, IEEE Transactions on Intelligent Transportation Systems.

[32]  Takeo Kanade,et al.  Accurate camera calibration using iterative refinement of control points , 2009, 2009 IEEE 12th International Conference on Computer Vision Workshops, ICCV Workshops.

[33]  Joris Heyman,et al.  TracTrac: A fast multi-object tracking algorithm for motion estimation , 2019, Comput. Geosci..

[34]  Wolfram Burgard,et al.  A benchmark for the evaluation of RGB-D SLAM systems , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[35]  S. An Extended Kalman Filter Based Automatic Frequency Control Loop , 2003 .

[36]  Rafael Muñoz-Salinas,et al.  Speeded up detection of squared fiducial markers , 2018, Image Vis. Comput..

[37]  Juan Carlos Niebles,et al.  D3TW: Discriminative Differentiable Dynamic Time Warping for Weakly Supervised Action Alignment and Segmentation , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[38]  Xinyan Deng,et al.  Flappy Hummingbird: An Open Source Dynamic Simulation of Flapping Wing Robots and Animals , 2019, 2019 International Conference on Robotics and Automation (ICRA).

[39]  Moritz Kampelmühler,et al.  Camera-based vehicle velocity estimation from monocular video , 2018, ArXiv.

[40]  Nasser Kehtarnavaz,et al.  UTD-MHAD: A multimodal dataset for human action recognition utilizing a depth camera and a wearable inertial sensor , 2015, 2015 IEEE International Conference on Image Processing (ICIP).

[41]  Kishore Mehrotra,et al.  Mixed coordinate tracking of generalized maneuvering targets using acceleration and jerk models , 2000, IEEE Trans. Aerosp. Electron. Syst..

[42]  Tarek Hamel,et al.  Attitude, Linear Velocity and Depth Estimation of a Camera Observing a Planar Target Using Continuous Homography and Inertial Data , 2018, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[43]  Dan Zecha,et al.  Activity-Conditioned Continuous Human Pose Estimation for Performance Analysis of Athletes Using the Example of Swimming , 2018, 2018 IEEE Winter Conference on Applications of Computer Vision (WACV).