Stopwords Detection in Bag-of-Visual-Words: The Case of Retrieving Maya Hieroglyphs

We present a method for automatic detection of stopwords in visual vocabularies that is based upon the entropy of each visual word. We propose a specific formulation to compute the entropy as the core of this method, in which the probability density function of the visual words is marginalized over all visual classes, such that words with higher entropy can be considered to be irrelevant words, i.e., stopwords. We evaluate our method on a dataset of syllabic Maya hieroglyphs, which is of great interest for archaeologists, and that requires efficient techniques for indexing and retrieval. Our results show that our method produces shorter bag representations without hurting retrieval performance, and even improving it in some cases, which does not happen when using previous methods. Furthermore, our assumptions for the proposed computation of the entropy can be generalized to bag representations of different nature.

[1]  Jean-Marc Odobez,et al.  Analyzing Ancient Maya Glyph Collections with Contextual Shape Descriptors , 2011, International Journal of Computer Vision.

[2]  Chong-Wah Ngo,et al.  Representations of Keypoint-Based Semantic Concept Detection: A Comprehensive Study , 2010, IEEE Transactions on Multimedia.

[3]  Matthieu Guillaumin,et al.  Segmentation Propagation in ImageNet , 2012, ECCV.

[4]  Zhipeng Zhao,et al.  Towards a local-global visual feature-based framework for recognition , 2009 .

[5]  Xiaotie Deng,et al.  Automatic construction of Chinese stop word list , 2006 .

[6]  Peter Ingwersen,et al.  Developing a Test Collection for the Evaluation of Integrated Search , 2010, ECIR.

[7]  Lei Zheng,et al.  Entropy-Based Static Index Pruning , 2009, ECIR.

[8]  Chong-Wah Ngo,et al.  Evaluating bag-of-visual-words representations in scene classification , 2007, MIR '07.

[9]  Cordelia Schmid,et al.  Aggregating local descriptors into a compact image representation , 2010, 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[10]  Roelof van Zwol,et al.  Spatially-aware indexing for image object retrieval , 2012, WSDM '12.

[11]  Alexander Hauptmann,et al.  Categorization Approach to Video Scene Classification Using Keypoint Features , 2006 .

[12]  David W. Corne,et al.  Towards modernised and Web-specific stoplists for Web document analysis , 2003, Proceedings IEEE/WIC International Conference on Web Intelligence (WI 2003).

[13]  Hervé Jégou,et al.  Negative Evidences and Co-occurences in Image Retrieval: The Benefit of PCA and Whitening , 2012, ECCV.

[14]  Andrew Zisserman,et al.  Video Google: a text retrieval approach to object matching in videos , 2003, Proceedings Ninth IEEE International Conference on Computer Vision.

[15]  Sang Joon Kim,et al.  A Mathematical Theory of Communication , 2006 .

[16]  Ming-Syan Chen,et al.  A Novel Language-Model-Based Approach for Image Object Mining and Re-ranking , 2008, 2008 Eighth IEEE International Conference on Data Mining.

[17]  Jean-Marc Odobez,et al.  A Thousand Words in a Scene , 2007, IEEE Transactions on Pattern Analysis and Machine Intelligence.