Optical-flow based self-supervised learning of obstacle appearance applied to MAV landing

Monocular optical flow has been widely used to detect obstacles in Micro Air Vehicles (MAVs) during visual navigation. However, this approach requires significant movement, which reduces the efficiency of navigation and may even introduce risks in narrow spaces. In this paper, we introduce a novel setup of self-supervised learning (SSL), in which optical flow cues serve as a scaffold to learn the visual appearance of obstacles in the environment. We apply it to a landing task, in which initially 'surface roughness' is estimated from the optical flow field in order to detect obstacles. Subsequently, a linear regression function is learned that maps appearance features represented by texton distributions to the roughness estimate. After learning, the MAV can detect obstacles by just analyzing a still image. This allows the MAV to search for a landing spot without moving. We first demonstrate this principle to work with offline tests involving images captured from an on-board camera, and then demonstrate the principle in flight. Although surface roughness is a property of the entire flow field in the global image, the appearance learning even allows for the pixel-wise segmentation of obstacles.

[1]  Davide Scaramuzza,et al.  SVO: Fast semi-direct monocular visual odometry , 2014, 2014 IEEE International Conference on Robotics and Automation (ICRA).

[2]  Matthew A. Garratt,et al.  Monocular Snapshot-based Sensing and Control of Hover, Takeoff, and Landing for a Low-cost Quadrotor , 2015, J. Field Robotics.

[3]  Roland Siegwart,et al.  Vision based MAV navigation in unknown and unstructured environments , 2010, 2010 IEEE International Conference on Robotics and Automation.

[4]  G. C. H. E. de Croon,et al.  Paparazzi: how to make a swarm of Parrot AR Drones fly autonomously based on GPS. , 2013 .

[5]  Guido C. H. E. de Croon,et al.  The appearance variation cue for obstacle avoidance , 2012, 2010 IEEE International Conference on Robotics and Biomimetics.

[6]  Robert E. Mahony,et al.  The landing problem of a VTOL Unmanned Aerial Vehicle on a moving platform using optical flow , 2010, 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[7]  Andrew Zisserman,et al.  Texture classification: are filter banks necessary? , 2003, 2003 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2003. Proceedings..

[8]  D. Izzo,et al.  Landing with Time-to-Contact and Ventral Optic Flow Estimates , 2012 .

[9]  Daniel Cremers,et al.  Semi-dense Visual Odometry for a Monocular Camera , 2013, 2013 IEEE International Conference on Computer Vision.

[10]  H. C. Longuet-Higgins,et al.  The interpretation of a moving retinal image , 1980, Proceedings of the Royal Society of London. Series B. Biological Sciences.

[11]  Nathan Michael,et al.  Vision-based Landing Site Evaluation and Trajectory Generation Toward Rooftop Landing , 2014, Robotics: Science and Systems.

[12]  H. W. Ho,et al.  Optical flow for self-supervised learning of obstacle appearance , 2015, 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[13]  Roland Brockers,et al.  Micro air vehicle autonomous obstacle avoidance from stereo-vision , 2014, Defense + Security Symposium.

[14]  Larry H. Matthies,et al.  Towards Autonomous Navigation of Miniature UAV , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition Workshops.

[15]  Giulio Reina,et al.  A Self‐learning Framework for Statistical Ground Classification using Radar and Monocular Vision , 2015, J. Field Robotics.

[16]  Robert C. Bolles,et al.  Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography , 1981, CACM.

[17]  P. Zingaretti,et al.  Autonomous safe landing of a vision guided helicopter , 2010, Proceedings of 2010 IEEE/ASME International Conference on Mechatronic and Embedded Systems and Applications.

[18]  Larry Matthies,et al.  Stereo vision and rover navigation software for planetary exploration , 2002, Proceedings, IEEE Aerospace Conference.

[19]  Mandyam V. Srinivasan,et al.  Visual Control of Flight Speed and Height in the Honeybee , 2006, SAB.

[20]  Peter H. N. de With,et al.  Free-Space Detection with Self-Supervised and Online Trained Fully Convolutional Networks , 2016, Autonomous Vehicles and Machines.

[21]  Soon-Jo Chung,et al.  MVCSLAM: Mono-Vision Corner SLAM for Autonomous Micro-Helicopters in GPS Denied Environments , 2008 .

[22]  Teuvo Kohonen,et al.  The self-organizing map , 1990 .

[23]  Sebastian Thrun,et al.  Reverse Optical Flow for Self-Supervised Adaptive Autonomous Robot Navigation , 2007, International Journal of Computer Vision.

[24]  Olivier Stasse,et al.  MonoSLAM: Real-Time Single Camera SLAM , 2007, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[25]  Stéphane Viollet,et al.  Neuromimetic Robots Inspired by Insect Vision , 2008 .

[26]  Urs A. Muller,et al.  Real-time adaptive off-road vehicle navigation and terrain classification , 2013, Defense, Security, and Sensing.

[27]  Andrew J. Davison,et al.  DTAM: Dense tracking and mapping in real-time , 2011, 2011 International Conference on Computer Vision.

[28]  Sang Joon Kim,et al.  A Mathematical Theory of Communication , 2006 .

[29]  Nicholas Roy,et al.  Autonomous Flight in Unknown Indoor Environments , 2009 .

[30]  Peter I. Corke,et al.  Two Seconds to Touchdown - Vision-Based Controlled Forced Landing , 2006, 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[31]  Xiaomin Wu,et al.  Self-paced cross-modality transfer learning for efficient road segmentation , 2017, 2017 IEEE International Conference on Robotics and Automation (ICRA).

[32]  B. Remes,et al.  Optic-Flow Based Slope Estimation for Autonomous Landing , 2013 .

[33]  Heinrich H. Bülthoff,et al.  Robust optical-flow based self-motion estimation for a quadrotor UAV , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[34]  Teuvo Kohonen,et al.  Self-Organizing Maps , 2010 .

[35]  Roland Siegwart,et al.  Monocular‐SLAM–based navigation for autonomous micro helicopters in GPS‐denied environments , 2011, J. Field Robotics.

[36]  Claude E. Shannon,et al.  The mathematical theory of communication , 1950 .

[37]  Daniel D. Lee,et al.  Online self-supervised monocular visual odometry for ground vehicles , 2015, 2015 IEEE International Conference on Robotics and Automation (ICRA).

[38]  Robert P. W. Duin,et al.  A Matlab Toolbox for Pattern Recognition , 2004 .

[39]  J.-Y. Bouguet,et al.  Pyramidal implementation of the lucas kanade feature tracker , 1999 .

[40]  Sebastian Thrun,et al.  Stanley: The robot that won the DARPA Grand Challenge , 2006, J. Field Robotics.

[41]  F. Dellaert,et al.  Optical Flow Templates for Superpixel Labeling in Autonomous Robot Navigation , 2013 .

[42]  Tom Drummond,et al.  Fusing points and lines for high performance tracking , 2005, Tenth IEEE International Conference on Computer Vision (ICCV'05) Volume 1.

[43]  Daniel Cremers,et al.  Semi-dense visual odometry for AR on a smartphone , 2014, 2014 IEEE International Symposium on Mixed and Augmented Reality (ISMAR).

[44]  J. M. M. Montiel,et al.  ORB-SLAM: A Versatile and Accurate Monocular SLAM System , 2015, IEEE Transactions on Robotics.

[45]  B. Remes,et al.  Design, Aerodynamics, and Vision-Based Control of the DelFly , 2009 .

[46]  Tom Drummond,et al.  Machine Learning for High-Speed Corner Detection , 2006, ECCV.

[47]  Murat Bronz,et al.  Using the Paparazzi UAV System for Scientific Research , 2014 .

[48]  Martial Hebert,et al.  Vision and Learning for Deliberative Monocular Cluttered Flight , 2014, FSR.

[49]  Anastasios I. Mourikis,et al.  High-precision, consistent EKF-based visual-inertial odometry , 2013, Int. J. Robotics Res..

[50]  Youdan Kim,et al.  Landing Site Searching Algorithm of a Quadrotor Using Depth Map of Stereo Vision on Unknown Terrain , 2012, Infotech@Aerospace.

[51]  Daniel Cremers,et al.  LSD-SLAM: Large-Scale Direct Monocular SLAM , 2014, ECCV.

[52]  Pascual Campoy Cervera,et al.  An Approach Toward Visual Autonomous Ship Board Landing of a VTOL UAV , 2014, 2013 International Conference on Unmanned Aircraft Systems (ICUAS).

[53]  Nicholas Roy,et al.  Stereo vision and laser odometry for autonomous helicopters in GPS-denied indoor environments , 2009, Defense + Commercial Sensing.

[54]  M. Srinivasan,et al.  Visual control of flight speed in honeybees , 2005, Journal of Experimental Biology.

[55]  Sebastian Thrun,et al.  Self-supervised Monocular Road Detection in Desert Terrain , 2006, Robotics: Science and Systems.

[56]  Norbert Boeddeker,et al.  A universal strategy for visually guided landing , 2013, Proceedings of the National Academy of Sciences.

[57]  Ingmar Posner,et al.  Find your own way: Weakly-supervised segmentation of path proposals for urban autonomy , 2016, 2017 IEEE International Conference on Robotics and Automation (ICRA).

[58]  Nicolas H. Franceschini,et al.  Aerial robot piloted in steep relief by optic flow sensors , 2008, 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems.