Transferring Learning To Rank Models for Web Search

Learning to rank techniques provide mechanisms for combining document feature values into learned models that produce effective rankings. However, issues concerning the transferability of learned models between different corpora or subsets of the same corpus are not yet well understood. For instance, is the importance of different feature sets consistent between subsets of a corpus, or whether a learned model obtained on a small subset of the corpus effectively transfer to the larger corpus? By formulating our experiments around two null hypotheses, in this work, we apply a full-factorial experiment design to empirically investigate these questions using the ClueWeb09 and ClueWeb12 corpora, combined with queries from the TREC Web track. Among other observations, our experiments reveal that Clue-Web09 remains an effective choice of training corpus for learning effective models for ClueWeb12, and also that the importance of query independent features varies among the ClueWeb09 and ClueWeb12 corpora. In doing so, this work contributes an important study into the transferability of learning to rank models, as well as empirically-derived best practices for effective retrieval on the ClueWeb12 corpus.

[1]  Shotaro Akaho,et al.  TrBagg: A Simple Transfer Learning Method and its Application to Personalization in Collaborative Tagging , 2009, 2009 Ninth IEEE International Conference on Data Mining.

[2]  Sinno Jialin Pan Transfer learning with applications on text, sensors and images , 2013, MLIS '13.

[3]  Craig MacDonald,et al.  Effectiveness beyond the first crawl tier , 2011, CIKM '11.

[4]  Tie-Yan Liu,et al.  Learning to rank for information retrieval , 2009, SIGIR.

[5]  Craig MacDonald,et al.  From Puppy to Maturity: Experiences in Developing Terrier , 2012, OSIR@SIGIR.

[6]  Rodrygo L. T. Santos,et al.  The whens and hows of learning to rank for web search , 2012, Information Retrieval.

[7]  Iadh Ounis,et al.  Incorporating term dependency in the dfr framework , 2007, SIGIR.

[8]  Tao Qin,et al.  LETOR: Benchmark Dataset for Research on Learning to Rank for Information Retrieval , 2007 .

[9]  W. Bruce Croft,et al.  Two-Stage Learning to Rank for Information Retrieval , 2013, ECIR.

[10]  Stephen Tyree,et al.  Parallel boosted regression trees for web search ranking , 2011, WWW.

[11]  W. Bruce Croft,et al.  A Markov random field model for term dependencies , 2005, SIGIR '05.

[12]  T. Minka Selection bias in the LETOR datasets , 2008 .

[13]  C. D. Gelatt,et al.  Optimization by Simulated Annealing , 1983, Science.

[14]  W. Bruce Croft,et al.  Quality-biased ranking of web documents , 2011, WSDM '11.

[15]  Qiang Yang,et al.  Boosting for transfer learning , 2007, ICML '07.

[16]  Eric R. Ziegel,et al.  The Elements of Statistical Learning , 2003, Technometrics.

[17]  Donald Metzler,et al.  Automatic feature selection in the markov random field model for information retrieval , 2007, CIKM '07.

[18]  Stephen E. Robertson,et al.  Okapi at TREC-3 , 1994, TREC.

[19]  Bahar Karaoglan,et al.  A nonparametric term weighting method for information retrieval based on measuring the divergence from independence , 2014, Information Retrieval.

[20]  Charles L. A. Clarke,et al.  Efficient and effective spam filtering and re-ranking for large web datasets , 2010, Information Retrieval.

[21]  Gilles Louppe,et al.  Learning to rank with extremely randomized trees , 2010, Yahoo! Learning to Rank Challenge.

[22]  Craig MacDonald,et al.  University of Glasgow at WebCLEF 2005: Experiments in per-field Normalisation and Language Specific Stemming , 2005, CLEF.

[23]  Ellen M. Voorhees,et al.  TREC 2014 Web Track Overview , 2015, TREC.

[24]  David A. Hull Using statistical testing in the evaluation of retrieval experiments , 1993, SIGIR.

[25]  Charles L. A. Clarke,et al.  Overview of the TREC 2012 Web Track , 2012, TREC.

[26]  Marc Najork,et al.  Breadth-first crawling yields high-quality pages , 2001, WWW '01.

[27]  Qiang Yang,et al.  A Survey on Transfer Learning , 2010, IEEE Transactions on Knowledge and Data Engineering.

[28]  Stephen E. Robertson,et al.  Microsoft Research at TREC 2009: Web and Relevance Feedback Track , 2009, TREC.

[29]  Christopher J. C. Burges,et al.  From RankNet to LambdaRank to LambdaMART: An Overview , 2010 .

[30]  Giorgio Gambosi,et al.  FUB, IASI-CNR and University of Tor Vergata at TREC 2008 Blog Track , 2008, TREC.

[31]  Bahar Karaoglan,et al.  IRRA at TREC 2010: Index Term Weighting by Divergence From Independence Model , 2010, TREC.

[32]  Yi Chang,et al.  Yahoo! Learning to Rank Challenge Overview , 2010, Yahoo! Learning to Rank Challenge.

[33]  Craig MacDonald,et al.  About learning models with multiple query-dependent features , 2013, TOIS.

[34]  Ellen M. Voorhees,et al.  Overview of the TREC 2014 Web Track , 2017 .