Enabling ∼100fps detection on a landing unmanned aircraft for its on-ground vision-based recovery

In this paper, a deep learning inspired solution is proposed and developed to enable timeliness and practicality of the pre-existing ground stereo vision guidance system for flxed-wing UAVs' safe landing. Since the ground guidance prototype was restricted within applications due to its untimeliness, eventually the vision-based detection less than 15 fps (frame per second). Under such circumstances, we employ a regression based deep learning algorithm into automatic detection on the flying aircraft in the landing sequential images. The system architecture is upgraded so as to be compatible with the novel deep learning requests, and furthermore, annotated datasets are conducted to support training and testing of the regression-based learning detection algorithm. Experimental results validate that the detection attaches 100 fps or more while the localization accuracy is kept in the same level.

[1]  Ross B. Girshick,et al.  Fast R-CNN , 2015, 1504.08083.

[2]  Tianjiang Hu,et al.  Ground Stereo Vision-Based Navigation for Autonomous Take-off and Landing of UAVs: A Chan-Vese Model Approach , 2016 .

[3]  Kaiming He,et al.  Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[4]  Farid Kendoul,et al.  Survey of advances in guidance, navigation, and control of unmanned rotorcraft systems , 2012, J. Field Robotics.

[5]  Tony F. Chan,et al.  Active contours without edges , 2001, IEEE Trans. Image Process..

[6]  Vijay Kumar,et al.  Opportunities and challenges with autonomous micro aerial vehicles , 2012, Int. J. Robotics Res..

[7]  Trevor Darrell,et al.  Caffe: Convolutional Architecture for Fast Feature Embedding , 2014, ACM Multimedia.

[8]  Trevor Darrell,et al.  Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation , 2013, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[9]  Tianjiang Hu,et al.  Chan-Vese model based binocular visual object extraction for UAV autonomous take-off and landing , 2015, 2015 5th International Conference on Information Science and Technology (ICIST).

[10]  Michael S. Lew,et al.  Deep learning for visual understanding: A review , 2016, Neurocomputing.

[11]  Fei-Fei Li,et al.  Large-Scale Video Classification with Convolutional Neural Networks , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[12]  Jianwei Zhang,et al.  Autonomous landing of an UAV with a ground-based actuated infrared stereo vision system , 2013, 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[13]  Daibing Zhang,et al.  Ground-based visual guidance in autonomous UAV landing , 2013, Other Conferences.

[14]  Zdenek Kalal,et al.  Tracking-Learning-Detection , 2012, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[15]  Silvio Savarese,et al.  Learning to Track at 100 FPS with Deep Regression Networks , 2016, ECCV.

[16]  Ali Farhadi,et al.  You Only Look Once: Unified, Real-Time Object Detection , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[17]  Tianjiang Hu,et al.  ROS-based ground stereo vision detection: implementation and experiments , 2016, Robotics and biomimetics.

[18]  Tianjiang Hu,et al.  Stereo Vision Guiding for the Autonomous Landing of Fixed-Wing UAVs: A Saliency-Inspired Approach , 2016 .

[19]  Shen Lincheng,et al.  A new calibration method for vision system using differential GPS , 2014, 2014 13th International Conference on Control Automation Robotics & Vision (ICARCV).