Aggressive navigation using high-speed natural feature point tracking

Presently, most autonomous aerial vehicles rely on satellite navigation such as GPS to sense their position in the earth reference frame. However, reliance on GPS restricts the vehicle to missions where GPS signals are readily received. Motion capture systems are capable of indoor localization but require large infrastructure and are prone to occlusion. To overcome these restrictions, a self-contained high-speed vision system was developed at the University of Tokyo in collaboration with Boeing Research & Technology. The system has been flight tested and shown to be capable of drift-free position and attitude estimates without any reliance on GPS signals. Furthermore, the positional accuracy and update rate is at least one order of magnitude superior to that of uncorrected GPS. The vision system combines a high-speed camera with a lightweight computer and power supply into a self-contained computer-vision package. The computer processes the incoming image stream with a modified version of the University of Oxford Parallel Tracking and Mapping (PTAM) SLAM algorithm. Using this algorithm, the location and pose of the camera (and the MA V it is attached to) is estimated as it moves through space by mapping natural features as they first appear and by tracking those features as they move through or reappear in the camera's view. Our vision system was demonstrated using a hexacopter test bed. In a pair of experiments, the hexacopter was able to autonomously repeat a circuit of takeoffs and landings at predetermined separated sites using only MEMs gyro sensors and our vision system. One experiment was performed inside Boeing's motioncapture equipped Collaborative Systems Laboratory (CSL) to prove independence from GPS and to measure the accuracy of the vision system. The hexacopter performed 5 circuits of the navigation task over an area approximately 8 x 8 m at an altitude of approximately 2 m. The vision system's camera was set to provide an image stream of 640 x 480 pixel resolution at 50 Hz. Upon comparison with motion capture data, position estimates from the vision system were shown to be free of drift, with an average error of 2.2 cm and a maximum error of 9.7 cm when the vision system coordinate frame was optimally aligned to the motion capture coordinate frame. A second experiment was performed in an open outdoor area, allowing for safe execution of more aggressive maneuvering. In this experiment, the vision system's camera was set to provide an image stream of 320 x 240 pixel resolution at 120 Hz. This experiment demonstrated the ability to perform takeoffs, landings and transit at higher speeds than was demonstrated in the indoor experiment.

[1]  Ian D. Reid,et al.  Real-Time SLAM Relocalisation , 2007, 2007 IEEE 11th International Conference on Computer Vision.

[2]  Vijay Kumar,et al.  Trajectory design and control for aggressive formation flight with quadrotors , 2012, Auton. Robots.

[3]  Bernhard P. Wrobel,et al.  Multiple View Geometry in Computer Vision , 2001 .

[4]  Suya You,et al.  Natural Feature Tracking for Augmented Reality , 1999, IEEE Trans. Multim..

[5]  David W. Murray,et al.  Parallel Tracking and Mapping on a camera phone , 2009, 2009 8th IEEE International Symposium on Mixed and Augmented Reality.

[6]  Éric Marchand,et al.  Accurate real-time tracking using mutual information , 2010, 2010 IEEE International Symposium on Mixed and Augmented Reality.

[7]  Tom Drummond,et al.  Machine Learning for High-Speed Corner Detection , 2006, ECCV.

[8]  Stefan Bieniawski,et al.  Vehicle Swarm Rapid Prototyping Testbed , 2009 .

[9]  G. Klein,et al.  Parallel Tracking and Mapping for Small AR Workspaces , 2007, 2007 6th IEEE and ACM International Symposium on Mixed and Augmented Reality.

[10]  Andrew J. Davison,et al.  Real-time simultaneous localisation and mapping with a single camera , 2003, Proceedings Ninth IEEE International Conference on Computer Vision.

[11]  David Nister,et al.  Recent developments on direct relative orientation , 2006 .

[12]  B. Ripley,et al.  Robust Statistics , 2018, Encyclopedia of Mathematical Geosciences.

[13]  George Loizou,et al.  Computer vision and pattern recognition , 2007, Int. J. Comput. Math..

[14]  Carlo Tomasi,et al.  Good features to track , 1994, 1994 Proceedings of IEEE Conference on Computer Vision and Pattern Recognition.