Our objective is to estimate the relevance of an image to a query for image search purposes. We address two limitations of the existing image search engines in this paper. First, there is no straightforward way of bridging the gap between semantic textual queries as well as users' search intents and image visual content. Image search engines therefore primarily rely on static and textual features. Visual features are mainly used to identify potentially useful recurrent patterns or relevant training examples for complementing search by image reranking. Second, image rankers are trained on query-image pairs labeled by human experts, making the annotation intellectually expensive and time-consuming. Furthermore, the labels may be subjective when the queries are ambiguous, resulting in difficulty in predicting the search intention. We demonstrate that the aforementioned two problems can be mitigated by exploring the use of click-through data, which can be viewed as the footprints of user searching behavior, as an effective means of understanding query. The correspondences between an image and a query are determined by whether the image was searched and clicked by users under the query in a commercial image search engine. We therefore hypothesize that the image click counts in response to a query are as their relevance indications. For each new image, our proposed graph-based label propagation algorithm employs neighborhood graph search to find the nearest neighbors on an image similarity graph built up with visual representations from deep neural networks and further aggregates their clicked queries/click counts to get the labels of the new image. We conduct experiments on MSR-Bing Grand Challenge and the results show consistent performance gain over various baselines. In addition, the proposed approach is very efficient, completing annotation of each query-image pair within just 15 milliseconds on a regular PC.
[1]
Shipeng Li,et al.
Query-driven iterated neighborhood graph search for large scale indexing
,
2012,
ACM Multimedia.
[2]
Nick Craswell,et al.
Random walks on the click graph
,
2007,
SIGIR.
[3]
Vidit Jain,et al.
Learning to re-rank: query-dependent image re-ranking using click data
,
2011,
WWW.
[4]
Chong-Wah Ngo,et al.
Annotation for free: video tagging by mining user search behavior
,
2013,
ACM Multimedia.
[5]
Ricardo Baeza-Yates,et al.
Query-sets: using implicit feedback and query patterns to organize web documents
,
2008,
WWW.
[6]
Luca Chiarandini,et al.
Image ranking based on user browsing behavior
,
2012,
SIGIR '12.
[7]
Jing Wang,et al.
Clickage: towards bridging semantic and intent gaps via mining click logs of search engines
,
2013,
ACM Multimedia.
[8]
Ha Hong,et al.
The Neural Representation Benchmark and its Evaluation on Brain and Machine
,
2013,
ICLR.
[9]
Ricardo A. Baeza-Yates,et al.
Extracting semantic relations from query logs
,
2007,
KDD '07.
[10]
Filip Radlinski,et al.
Evaluating the accuracy of implicit feedback from clicks and query reformulations in Web search
,
2007,
TOIS.
[11]
Xiao Li,et al.
Learning query intent from regularized click graphs
,
2008,
SIGIR '08.
[12]
Qi Tian,et al.
Multimedia search reranking: A literature survey
,
2014,
CSUR.