Multi-Camera Tracking and Mapping for Unmanned Aerial Vehicles in Unstructured Environments

Pose estimation for small unmanned aerial vehicles has made large improvements in recent years, leading to vehicles that use a suite of sensors to navigate and explore various environments. In particular, cameras have become popular due to their low weight and power consumption, as well as the large amount of data they capture. However, processing this data to extract useful information has proved challenging, as the pose estimation problem is inherently nonlinear and, depending on the cameras’ field of view, potentially ill-posed. Results from the field of multi-camera egomotion estimation show that these issues can be reduced or eliminated by using multiple cameras positioned appropriately. In this work, we make use of these insights to develop a multi-camera visual pose estimator using ultra wide angle fisheye cameras, leading to a system that has many advantages over traditional visual pose estimators. The system is tested in a variety of configurations and flight scenarios on an unprepared urban rooftop, including landings and takeoffs. To our knowledge, this is the first time a visual pose estimator has been shown to be able to continuously track the pose of a small aerial vehicle throughout the landing and subsequent takeoff maneuvers.

[1]  Fadi Dornaika,et al.  Hand-Eye Calibration , 1995, Int. J. Robotics Res..

[2]  C. Fermuller,et al.  Eyes from eyes: new cameras for structure from motion , 2002, Proceedings of the IEEE Workshop on Omnidirectional Vision 2002. Held in conjunction with ECCV'02.

[3]  Robert Pless,et al.  A spherical eye from multiple cameras (makes better models of the world) , 2001, Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition. CVPR 2001.

[4]  Ian D. Reid,et al.  RSLAM: A System for Large-Scale Mapping in Constant-Time Using Stereo , 2011, International Journal of Computer Vision.

[5]  Frank Dellaert,et al.  Probabilistic structure matching for visual SLAM with a multi-camera rig , 2010, Comput. Vis. Image Underst..

[6]  Vijay Kumar,et al.  Trajectory Generation and Control for Precise Aggressive Maneuvers with Quadrotors , 2010, ISER.

[7]  Andrew W. Fitzgibbon,et al.  Bundle Adjustment - A Modern Synthesis , 1999, Workshop on Vision Algorithms.

[8]  Reinhard Koch,et al.  Calibration of a Multi-camera Rig from Non-overlapping Views , 2007, DAGM-Symposium.

[9]  Hauke Strasdat,et al.  Real-time monocular SLAM: Why filter? , 2010, 2010 IEEE International Conference on Robotics and Automation.

[10]  Davide Scaramuzza,et al.  Omnidirectional Vision: From Calibration to Root Motion Estimation , 2007 .

[11]  Olivier D. Faugeras,et al.  Automatic calibration and removal of distortion from scenes of structured environments , 1995, Optics & Photonics.

[12]  Quan Pan,et al.  Pose estimation for multi-camera systems , 2017, 2017 IEEE International Conference on Unmanned Systems (ICUS).

[13]  Luis Puig,et al.  Visual SLAM with an Omnidirectional Camera , 2010, 2010 20th International Conference on Pattern Recognition.

[14]  Roland Siegwart,et al.  Monocular‐SLAM–based navigation for autonomous micro helicopters in GPS‐denied environments , 2011, J. Field Robotics.

[15]  Metin Tarhan,et al.  Control of a quadrotor air vehicle by vanishing points in catadioptric images , 2009, 2009 International Symposium on Optomechatronic Technologies.

[16]  G. Klein,et al.  Parallel Tracking and Mapping for Small AR Workspaces , 2007, 2007 6th IEEE and ACM International Symposium on Mixed and Augmented Reality.

[17]  Alejandro Rituerto,et al.  ´Indice de Contenidos A Comparison of omnidirectional and conventional monocular systems for visual SLAM , 2010 .

[18]  Wolfram Burgard,et al.  G2o: A general framework for graph optimization , 2011, 2011 IEEE International Conference on Robotics and Automation.

[19]  Roland Siegwart,et al.  Intuitive 3D Maps for MAV Terrain Exploration and Obstacle Avoidance , 2011, J. Intell. Robotic Syst..

[20]  Roland Siegwart,et al.  Vision based MAV navigation in unknown and unstructured environments , 2010, 2010 IEEE International Conference on Robotics and Automation.

[21]  Adam Harmat,et al.  Parallel Tracking and Mapping with Multiple Cameras on an Unmanned Aerial Vehicle , 2012, ICIRA.

[22]  Olivier Stasse,et al.  MonoSLAM: Real-Time Single Camera SLAM , 2007, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[23]  Morgan Quigley,et al.  ROS: an open-source Robot Operating System , 2009, ICRA 2009.

[24]  Ian D. Reid,et al.  Adaptive relative bundle adjustment , 2009, Robotics: Science and Systems.

[25]  Jan-Michael Frahm,et al.  Robust 6DOF Motion Estimation for Non-Overlapping, Multi-Camera Systems , 2008, 2008 IEEE Workshop on Applications of Computer Vision.

[26]  Jan-Michael Frahm,et al.  Visual Odometry for Non-overlapping Views Using Second-Order Cone Programming , 2007, ACCV.

[27]  Teresa A. Vidal-Calleja,et al.  Fusing Monocular Information in Multicamera SLAM , 2008, IEEE Transactions on Robotics.

[28]  Wolfgang Förstner,et al.  Bundle Adjustment for Multi-Camera Systems with Points at Infinity , 2012 .

[29]  Roland Siegwart,et al.  Real-time 6D stereo Visual Odometry with non-overlapping fields of view , 2012, 2012 IEEE Conference on Computer Vision and Pattern Recognition.

[30]  Kurt Konolige,et al.  Double window optimisation for constant time visual SLAM , 2011, 2011 International Conference on Computer Vision.

[31]  Takeo Kanade,et al.  Motion estimation using multiple non-overlapping cameras for small unmanned aerial vehicles , 2008, 2008 IEEE International Conference on Robotics and Automation.

[32]  Marc Pollefeys,et al.  Vision-based autonomous mapping and exploration using a quadrotor MAV , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[33]  Andrew J. Davison,et al.  SLAM-based automatic extrinsic calibration of a multi-camera rig , 2011, 2011 IEEE International Conference on Robotics and Automation.

[34]  M. Rubin,et al.  Arithmetic and geometric solutions for average rigid-body rotation , 2010 .

[35]  Yiannis Aloimonos,et al.  Observability of 3D Motion , 2000, International Journal of Computer Vision.

[36]  K. H. Wong,et al.  Multiple nonoverlapping camera pose estimation , 2010, 2010 IEEE International Conference on Image Processing.

[37]  Frank Dellaert,et al.  Square Root SAM: Simultaneous Localization and Mapping via Square Root Information Smoothing , 2006, Int. J. Robotics Res..

[38]  Andreas Zell,et al.  On-board dual-stereo-vision for autonomous quadrotor navigation , 2013, 2013 International Conference on Unmanned Aircraft Systems (ICUAS).

[39]  Jean-Arcady Meyer,et al.  Real-time visual loop-closure detection , 2008, 2008 IEEE International Conference on Robotics and Automation.

[40]  Robert Pless,et al.  Using many cameras as one , 2003, 2003 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2003. Proceedings..

[41]  Andrew J. Davison,et al.  Real-Time Camera Tracking: When is High Frame-Rate Best? , 2012, ECCV.

[42]  David W. Murray,et al.  Improving the Agility of Keyframe-Based SLAM , 2008, ECCV.

[43]  Jonathan P. How,et al.  Vision-based guidance and control of a hovering vehicle in unknown, GPS-denied environments , 2009, 2009 IEEE International Conference on Robotics and Automation.

[44]  Pascal Vasseur,et al.  Omnidirectional vision on UAV for attitude computation , 2006, Proceedings 2006 IEEE International Conference on Robotics and Automation, 2006. ICRA 2006..

[45]  Cyrill Stachniss,et al.  Hierarchical optimization on manifolds for online 2D and 3D mapping , 2010, 2010 IEEE International Conference on Robotics and Automation.