Evaluation of visual object retrieval datasets

Recently visual object retrieval has being widely studied for its vast application prospect, and the progress is usually tracked by the performance achieved on benchmark datasets. Therefore, in order to faithfully evaluate object retrieval algorithms, the quality of datasets must be ensured by some means. In this paper, we propose a method to evaluate the quality of benchmarking visual object retrieval on two highly cited object retrieval datasets, the Oxford datasets and TrecVid instance search datasets. Our evaluation method leverages the essential differences between object retrieval and other similar image search, and digs out some unrevealed and rather interesting features from those datasets. To the best of our knowledge, this research has never been touched before. Our method is believed to be beneficial to dataset collection and algorithm evaluation of object retrieval. More importantly, we hope this work can attract more attentions on this topic in the community.

[1]  Andrew Zisserman,et al.  Video Google: a text retrieval approach to object matching in videos , 2003, Proceedings Ninth IEEE International Conference on Computer Vision.

[2]  David Nistér,et al.  Scalable Recognition with a Vocabulary Tree , 2006, 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06).

[3]  Georges Quénot,et al.  TRECVID 2015 - An Overview of the Goals, Tasks, Data, Evaluation Mechanisms and Metrics , 2011, TRECVID.

[4]  Cordelia Schmid,et al.  Scale & Affine Invariant Interest Point Detectors , 2004, International Journal of Computer Vision.

[5]  Christoph H. Lampert Detecting objects in large image collections and videos by efficient subimage retrieval , 2009, 2009 IEEE 12th International Conference on Computer Vision.

[6]  Cordelia Schmid,et al.  Hamming Embedding and Weak Geometric Consistency for Large Scale Image Search , 2008, ECCV.

[7]  Michael Isard,et al.  Total Recall: Automatic Query Expansion with a Generative Feature Model for Object Retrieval , 2007, 2007 IEEE 11th International Conference on Computer Vision.

[8]  Shin'ichi Satoh,et al.  Large vocabulary quantization for searching instances from videos , 2012, ICMR '12.

[9]  Jiri Matas,et al.  Efficient representation of local geometry for large scale object retrieval , 2009, CVPR.

[10]  Brian Antonishek TRECVID 2010 – An Introduction to the Goals , Tasks , Data , Evaluation Mechanisms , and Metrics , 2010 .

[11]  Yuning Jiang,et al.  Grid-based local feature bundling for efficient object search and localization , 2011, 2011 18th IEEE International Conference on Image Processing.

[12]  Alexei A. Efros,et al.  Unbiased look at dataset bias , 2011, CVPR 2011.

[13]  G LoweDavid,et al.  Distinctive Image Features from Scale-Invariant Keypoints , 2004 .

[14]  Jiri Matas,et al.  Total recall II: Query expansion revisited , 2011, CVPR 2011.

[15]  Andrew Zisserman,et al.  Three things everyone should know to improve object retrieval , 2012, 2012 IEEE Conference on Computer Vision and Pattern Recognition.

[16]  Michael Isard,et al.  Object retrieval with large vocabularies and fast spatial matching , 2007, 2007 IEEE Conference on Computer Vision and Pattern Recognition.

[17]  Michael Isard,et al.  Lost in quantization: Improving particular object retrieval in large scale image databases , 2008, 2008 IEEE Conference on Computer Vision and Pattern Recognition.