Optical Flow for Collision Avoidance in Autonomous Cars

Autonomous cars, robotic platforms and other devices capable of unassisted movement are becoming widely considered as superior to human-based control in many areas. Such platforms, however, often are constructed using expensive equipment. We investigate the possibility of using simple setup consisting of an embedded computer module, such as smartphone, single camera and inertial measurement unit along with the concept of the optical flow to detect possible collisions given real-time, onboard processing as an alternative to compound systems based on radar and lidar devices. While most optical flow algorithms are not applicable for real-time processing, our findings prove that those which sacrifice accuracy to gain speed can still be upgraded while remaining accurate enough for the field of collision avoidance. We propose modifications to further enhance optical flow for given context. Our findings prove that using proposed setup consisting of both hardware and software allows for omitting expensive sensors in the field of collision avoidance.

[1]  Aurélien Plyer,et al.  A Prediction-Correction Approach for Real-Time Optical Flow Computation Using Stereo , 2016, GCPR.

[2]  Frédéric Champagnat,et al.  Dense optical flow by iterative local window registration , 2005, IEEE International Conference on Image Processing 2005.

[3]  Andreas Geiger,et al.  Object scene flow for autonomous vehicles , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[4]  Antonio González Muñoz,et al.  Fuzzy behaviors for mobile robot navigation: design, coordination and fusion , 2000, Int. J. Approx. Reason..

[5]  Sebastian Thrun,et al.  Toward robotic cars , 2010, CACM.

[6]  Agnieszka Szczęsna,et al.  Model-based extended quaternion Kalman filter to inertial orientation tracking of arbitrary kinematic chains , 2016, SpringerPlus.

[7]  Ewa Lach,et al.  Inertial Motion Capture Costume Design Study , 2017, Sensors.

[8]  Christian Heipke,et al.  Joint 3d Estimation of Vehicles and Scene Flow , 2015 .

[9]  Richard P. Wildes,et al.  Coarse-to-fine stereo vision with accurate 3D boundaries , 2010, Image Vis. Comput..

[10]  Zhengyou Zhang,et al.  A Flexible New Technique for Camera Calibration , 2000, IEEE Trans. Pattern Anal. Mach. Intell..

[11]  Directorate Resources Science at the service of Europe : a guide to the 2007 annual activity report of the European Commission's Directorate-General for Research , 2008 .

[12]  Francisco Bonin-Font,et al.  Visual Navigation for Mobile Robots: A Survey , 2008, J. Intell. Robotic Syst..

[13]  Avinash C. Kak,et al.  Vision for Mobile Robot Navigation: A Survey , 2002, IEEE Trans. Pattern Anal. Mach. Intell..

[14]  Andrzej Polanski,et al.  Novel Lightweight Quaternion Filter for Determining Orientation Based on Indications of Gyroscope, Magnetometer and Accelerometer , 2014, ICCVG.

[15]  Aurélien Plyer,et al.  Massively parallel Lucas Kanade optical flow for real-time video processing applications , 2014, Journal of Real-Time Image Processing.

[16]  Bingbing Ni,et al.  Unsupervised Deep Learning for Optical Flow Estimation , 2017, AAAI.

[17]  Guy Le Besnerais,et al.  eVO: A realtime embedded stereo odometry for MAV applications , 2013, 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems.