Robust Online Learning to Rank via Selective Pairwise Approach Based on Evaluation Measures

Learning to rank is a supervised learning problem whose goal is to construct a ranking model. In recent years, online learning to rank algorithms have begun to attract attention because large-scale datasets have become available. We propose a selective pairwise approach to online learning to rank algorithms that offer both fast learning and high performance. The basic strategy of our method is to select the most effective document pair to minimize the objective function using an entered query present in the training data, and then updates the current weight vector by using only the selected document pair instead of using all document pairs in the query. The main characteristics of our method are that it utilizes adaptive margin rescaling based on the approximated NDCG to reflect the IR evaluation measure, the max-loss update procedure, and ramp loss to reduce the over-fitting problem. Finally, we implement our proposal, PARank-NDCG, in the framework of the Passive-Aggressive algorithm. We conduct experiments on the MSLRWEB datasets, which contain 10,000 and 30,000 queries. Our experiments show that PARank-NDCG outperforms conventional algorithms including online learning to rank algorithms such as Stochastic Pairwise Descent, Committee Perceptron and batch algorithm such as RankingSVM on NDCG values. In addition, our method only takes 7 seconds to learn a model on the MSLR-WEB10K dataset. PARank-NDCG offers approximately 63 times faster training than RankingSVM on average.

[1]  Charles L. A. Clarke,et al.  Information Retrieval - Implementing and Evaluating Search Engines , 2010 .

[2]  D. Sculley,et al.  Large Scale Learning to Rank , 2009 .

[3]  Thorsten Joachims,et al.  Optimizing search engines using clickthrough data , 2002, KDD.

[4]  Slobodan Vucetic,et al.  Online Passive-Aggressive Algorithms on a Budget , 2010, AISTATS.

[5]  Zhaohui Zheng,et al.  Session Based Click Features for Recency Ranking , 2010, AAAI.

[6]  Quoc V. Le,et al.  Learning to Rank with Nonsmooth Cost Functions , 2006, Neural Information Processing Systems.

[7]  Aravind K. Joshi,et al.  Ranking and Reranking with Perceptron , 2005, Machine Learning.

[8]  Pawan Kumar,et al.  Notice of Violation of IEEE Publication Principles The Anatomy of a Large-Scale Hyper Textual Web Search Engine , 2009 .

[9]  Jason Weston,et al.  Trading convexity for scalability , 2006, ICML.

[10]  Sergey Brin,et al.  The Anatomy of a Large-Scale Hypertextual Web Search Engine , 1998, Comput. Networks.

[11]  Jaime G. Carbonell,et al.  Fast learning of document ranking functions with the committee perceptron , 2008, WSDM '08.

[12]  Stephen E. Robertson,et al.  SoftRank: optimizing non-smooth rank metrics , 2008, WSDM '08.

[13]  Jaana Kekäläinen,et al.  Cumulated gain-based evaluation of IR techniques , 2002, TOIS.

[14]  Tie-Yan Liu,et al.  Learning to rank for information retrieval , 2009, SIGIR.

[15]  Chiranjib Bhattacharyya,et al.  Structured learning for non-smooth ranking losses , 2008, KDD.

[16]  Thorsten Joachims,et al.  A support vector method for multivariate performance measures , 2005, ICML.

[17]  Wei Chu,et al.  Online learning for recency search ranking using real-time user feedback , 2010, CIKM '10.

[18]  Stephen E. Robertson,et al.  Okapi at TREC-3 , 1994, TREC.

[19]  Tie-Yan Liu,et al.  Future directions in learning to rank , 2010, Yahoo! Learning to Rank Challenge.

[20]  増山 繁,et al.  Online Passive-Aggressive Algorithmを用いたクラスタリング , 2010 .

[21]  Michael Collins,et al.  Discriminative Training Methods for Hidden Markov Models: Theory and Experiments with Perceptron Algorithms , 2002, EMNLP.

[22]  Tao Qin,et al.  LETOR: Benchmark Dataset for Research on Learning to Rank for Information Retrieval , 2007 .

[23]  Rong Jin,et al.  Learning to Rank by Optimizing NDCG Measure , 2009, NIPS.

[24]  Gilad Mishne,et al.  Towards recency ranking in web search , 2010, WSDM '10.

[25]  Koby Crammer,et al.  Pranking with Ranking , 2001, NIPS.

[26]  Filip Radlinski,et al.  A support vector method for optimizing average precision , 2007, SIGIR.

[27]  O. Chapelle Large margin optimization of ranking measures , 2007 .

[28]  Hang Li,et al.  AdaRank: a boosting algorithm for information retrieval , 2007, SIGIR.

[29]  Hang Li,et al.  Improving quality of training data for learning to rank using click-through data , 2010, WSDM '10.

[30]  Tie-Yan Liu,et al.  Adapting ranking SVM to document retrieval , 2006, SIGIR.

[31]  Koby Crammer,et al.  Online Passive-Aggressive Algorithms , 2003, J. Mach. Learn. Res..