An Online Learning Framework for Refining Recency Search Results with User Click Feedback

Traditional machine-learned ranking systems for Web search are often trained to capture stationary relevance of documents to queries, which have limited ability to track nonstationary user intention in a timely manner. In recency search, for instance, the relevance of documents to a query on breaking news often changes significantly over time, requiring effective adaptation to user intention. In this article, we focus on recency search and study a number of algorithms to improve ranking results by leveraging user click feedback. Our contributions are threefold. First, we use commercial search engine sessions collected in a random exploration bucket for reliable offline evaluation of these algorithms, which provides an unbiased comparison across algorithms without online bucket tests. Second, we propose an online learning approach that reranks and improves the search results for recency queries near real-time based on user clicks. This approach is very general and can be combined with sophisticated click models. Third, our empirical comparison of a dozen algorithms on real-world search data suggests importance of a few algorithmic choices in these applications, including generalization across different query-document pairs, specialization to popular queries, and near real-time adaptation of user clicks for reranking.

[1]  Mehryar Mohri,et al.  Magnitude-preserving ranking algorithms , 2007, ICML '07.

[2]  Zhaohui Zheng,et al.  Learning to re-rank web search results with multiple pairwise features , 2011, WSDM '11.

[3]  Thorsten Joachims,et al.  Interactively optimizing information retrieval systems as a dueling bandits problem , 2009, ICML '09.

[4]  Bee-Chung Chen,et al.  Explore/Exploit Schemes for Web Content Optimization , 2009, 2009 Ninth IEEE International Conference on Data Mining.

[5]  Gilad Mishne,et al.  Towards recency ranking in web search , 2010, WSDM '10.

[6]  Susan T. Dumais,et al.  Leveraging temporal dynamics of document content in relevance ranking , 2010, WSDM '10.

[7]  Wei Chu,et al.  Online learning for recency search ranking using real-time user feedback , 2010, CIKM '10.

[8]  Yehuda Koren,et al.  Collaborative filtering with temporal dynamics , 2009, KDD.

[9]  H. Robbins Some aspects of the sequential design of experiments , 1952 .

[10]  Fernando Diaz,et al.  Integration of news content into web results , 2009, WSDM '09.

[11]  Joaquin Quiñonero Candela,et al.  Web-Scale Bayesian Click-Through rate Prediction for Sponsored Search Advertising in Microsoft's Bing Search Engine , 2010, ICML.

[12]  Thore Graepel,et al.  Matchbox: large scale online bayesian recommendations , 2009, WWW '09.

[13]  Zheng Chen,et al.  A novel click model and its applications to online advertising , 2010, WSDM '10.

[14]  Thorsten Joachims,et al.  Optimizing search engines using clickthrough data , 2002, KDD.

[15]  Chao Liu,et al.  Click chain model in web search , 2009, WWW '09.

[16]  Quoc V. Le,et al.  Abstract , 2003, Appetite.

[17]  Gregory N. Hullender,et al.  Learning to rank using gradient descent , 2005, ICML.

[18]  Hongyuan Zha,et al.  A General Boosting Method and its Application to Learning Ranking Functions for Web Search , 2007, NIPS.

[19]  Wei Chu,et al.  Unbiased offline evaluation of contextual-bandit-based news article recommendation algorithms , 2010, WSDM '11.

[20]  Brian D. Davison,et al.  Freshness matters: in flowers, food, and web authority , 2010, SIGIR.

[21]  LiLihong,et al.  An Online Learning Framework for Refining Recency Search Results with User Click Feedback , 2012 .

[22]  Deepak Agarwal,et al.  Fast online learning through offline initialization for time-sensitive recommendation , 2010, KDD.

[23]  Wei Vivian Zhang,et al.  Comparing Click Logs and Editorial Labels for Training Query Rewriting , 2007 .

[24]  Hongyuan Zha,et al.  Global ranking by exploiting user clicks , 2009, SIGIR.

[25]  Ciya Liao,et al.  A model to estimate intrinsic document relevance from the clickthrough logs of a web search engine , 2010, WSDM '10.

[26]  Chao Liu,et al.  Post-rank reordering: resolving preference misalignments between search engines and end users , 2009, CIKM.

[27]  Fernando Diaz,et al.  Time is of the essence: improving recency ranking using Twitter data , 2010, WWW '10.

[28]  Yoram Singer,et al.  An Efficient Boosting Algorithm for Combining Preferences by , 2013 .

[29]  Olivier Chapelle,et al.  A dynamic bayesian network click model for web search ranking , 2009, WWW '09.

[30]  Filip Radlinski,et al.  How does clickthrough data reflect retrieval quality? , 2008, CIKM '08.

[31]  Zhaohui Zheng,et al.  Session Based Click Features for Recency Ranking , 2010, AAAI.

[32]  Filip Radlinski,et al.  Active exploration for learning rankings from clickthrough data , 2007, KDD '07.

[33]  Wei Chu,et al.  A contextual-bandit approach to personalized news article recommendation , 2010, WWW '10.

[34]  Jaana Kekäläinen,et al.  Cumulated gain-based evaluation of IR techniques , 2002, TOIS.

[35]  Tie-Yan Liu,et al.  Learning to rank for information retrieval , 2009, SIGIR.

[36]  Susan T. Dumais,et al.  Understanding temporal query dynamics , 2011, WSDM '11.

[37]  Fernando Diaz,et al.  Temporal profiles of queries , 2007, TOIS.