Performance Evaluation of Image Feature Detectors and Descriptors for Outdoor-Scene Visual Navigation

Scene image matching is often used for positioning of a visually navigated autonomous robot. The robot memorizes the scene as an image at each navigation point in the teaching mode, and knows being at the same position when the outside scene is matched to the image in the playback mode. The scene matching is usually accomplished by feature-based image matching methods, such as SIFT or SURF. However the problem is that matching results of such methods are greatly affected by changes in illumination condition. Therefore, it is important to know which method is robust to the illumination change. Several performance evaluation results of these matching methods have been reported, but they are not focusing on illumination change problem. In this paper, we present performance comparison results of these feature-based image matching methods against illumination change in outdoor scenes assuming usage for visual navigation purpose. We also encounter another problem when conducting such the comparison for visual navigation. In this application, the matching score gradually increases as approaching the matching point, and gradually decreases as being apart from that point. This impedes to define the right matching response (ground truth). We present one method by which giving the right response.

[1]  Masayuki Inaba,et al.  Visual navigation using view-sequenced route representation , 1996, Proceedings of IEEE International Conference on Robotics and Automation.

[2]  Ian Horswill,et al.  Polly: A Vision-Based Artificial Agent , 1993, AAAI.

[3]  Tom Fawcett,et al.  An introduction to ROC analysis , 2006, Pattern Recognit. Lett..

[4]  Krystian Mikolajczyk,et al.  Evaluation of local detectors and descriptors for fast feature matching , 2012, Proceedings of the 21st International Conference on Pattern Recognition (ICPR2012).

[5]  Cordelia Schmid,et al.  A Performance Evaluation of Local Descriptors , 2005, IEEE Trans. Pattern Anal. Mach. Intell..

[6]  Christopher G. Harris,et al.  A Combined Corner and Edge Detector , 1988, Alvey Vision Conference.

[7]  Jiri Matas,et al.  Robust wide-baseline stereo from maximally stable extremal regions , 2004, Image Vis. Comput..

[8]  David G. Lowe,et al.  Distinctive Image Features from Scale-Invariant Keypoints , 2004, International Journal of Computer Vision.

[9]  S. Yuta,et al.  Outdoor landmark map generation through human assisted route teaching for mobile robot navigation , 1999 .

[10]  Tamim Asfour,et al.  Combining Harris interest points and the SIFT descriptor for fast scale-invariant object recognition , 2009, 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[11]  Jean-Michel Morel,et al.  ASIFT: A New Framework for Fully Affine Invariant Image Comparison , 2009, SIAM J. Imaging Sci..

[12]  Matthijs C. Dorst Distinctive Image Features from Scale-Invariant Keypoints , 2011 .

[13]  Pietro Perona,et al.  Evaluation of Features Detectors and Descriptors based on 3D Objects , 2005, Tenth IEEE International Conference on Computer Vision (ICCV'05) Volume 1.

[14]  Luc Van Gool,et al.  SURF: Speeded Up Robust Features , 2006, ECCV.

[15]  Tom Drummond,et al.  Machine Learning for High-Speed Corner Detection , 2006, ECCV.

[16]  Akihisa Ohya,et al.  Outdoor landmark map generation through human route teaching for mobile robot navigation , 1996, Proceedings of IEEE/RSJ International Conference on Intelligent Robots and Systems. IROS '96.

[17]  Andrea Vedaldi,et al.  Vlfeat: an open and portable library of computer vision algorithms , 2010, ACM Multimedia.