ERR.Rank: An algorithm based on learning to rank for direct optimization of Expected Reciprocal Rank

Learning to rank (LTR) is a machine learning-based ranking technique that constructs a ranking model to sort objects in response to a query, and is used in many applications especially in information retrieval. LTR ranking models are generally evaluated using information retrieval measures. Listwise approaches are among the most important learning to rank algorithms. A subset of listwise approaches try to optimize the evaluation measures. These evaluation measures are dependent only on the document positions in the ranking and are discontinuous and non-convex with respect to the scores of the ranking function. The majority of evaluation measures used by current listwise techniques ignore the relationship between a document at a position and the documents at higher positions. To overcome this problem, we propose a new listwise algorithm, which aims to directly optimize the Expected Reciprocal Rank (ERR) measure. ERR considers the importance of a document at a position to be dependent on the documents ranked higher than this document. Our algorithm uses a probabilistic framework to optimize the expected value of ERR. We use a boosting approach using a gradient descent in order to find the optimal ranking function. The presented algorithm is compared with state of the art algorithms. The results obtained on the ’LETOR 3.0’ standard dataset indicate that the proposed method outperforms the baselines.

[1]  Mingrui Wu,et al.  Gradient descent optimization of smoothed information retrieval metrics , 2010, Information Retrieval.

[2]  Stephen E. Robertson,et al.  On the choice of effectiveness measures for learning to rank , 2010, Information Retrieval.

[3]  Tao Qin,et al.  Query-level loss functions for information retrieval , 2008, Inf. Process. Manag..

[4]  Tao Qin,et al.  A general approximation framework for direct optimization of information retrieval measures , 2010, Information Retrieval.

[5]  José Luis Vicedo González,et al.  TREC: Experiment and evaluation in information retrieval , 2007, J. Assoc. Inf. Sci. Technol..

[6]  Rossana M. de Castro Andrade,et al.  MLM-rank: A Ranking Algorithm Based on the Minimal Learning Machine , 2015, 2015 Brazilian Conference on Intelligent Systems (BRACIS).

[7]  Gregory N. Hullender,et al.  Learning to rank using gradient descent , 2005, ICML.

[8]  Xueqi Cheng,et al.  Directly Optimize Diversity Evaluation Measures , 2017, ACM Trans. Intell. Syst. Technol..

[9]  Falk Scholer,et al.  Effective Pre-retrieval Query Performance Prediction Using Similarity and Variability Evidence , 2008, ECIR.

[10]  Maksims Volkovs,et al.  BoltzRank: learning to maximize expected ranking gain , 2009, ICML '09.

[11]  Tie-Yan Liu,et al.  Directly optimizing evaluation measures in learning to rank , 2008, SIGIR.

[12]  Olivier Chapelle,et al.  Expected reciprocal rank for graded relevance , 2009, CIKM.

[13]  Ping Zhang,et al.  Learning to Rank by Optimizing Expected Reciprocal Rank , 2011, AIRS.

[14]  S. Sathiya Keerthi,et al.  Efficient algorithms for ranking with SVMs , 2010, Information Retrieval.

[15]  Martha Larson,et al.  CLiMF: learning to maximize reciprocal rank with collaborative less-is-more filtering , 2012, RecSys.

[16]  Filip Radlinski,et al.  A support vector method for optimizing average precision , 2007, SIGIR.

[17]  Hang Li,et al.  AdaRank: a boosting algorithm for information retrieval , 2007, SIGIR.

[18]  Wei-Ying Ma,et al.  Direct Optimization of Evaluation Measures in Learning to Rank , 2010 .

[19]  Tie-Yan Liu,et al.  Learning to Rank for Information Retrieval , 2011 .

[20]  Djoerd Hiemstra,et al.  Ranking Learning-to-Rank Methods , 2017, LEARNER@ICTIR.

[21]  Chiranjib Bhattacharyya,et al.  Structured learning for non-smooth ranking losses , 2008, KDD.

[22]  Ricardo da Silva Torres,et al.  On interactive learning-to-rank for IR: Overview, recent advances, challenges, and directions , 2016, Neurocomputing.

[23]  Charles L. A. Clarke,et al.  Overview of the TREC 2004 Terabyte Track , 2004, TREC.

[24]  Tie-Yan Liu,et al.  Learning to rank: from pairwise approach to listwise approach , 2007, ICML '07.

[25]  Qiang Wu,et al.  Adapting boosting for information retrieval measures , 2010, Information Retrieval.

[26]  Ben He,et al.  Direct measurement of training query quality for learning to rank , 2016, SAC.

[27]  Ambuj Tewari,et al.  Generalization error bounds for learning to rank: Does the length of document lists matter? , 2015, ICML.

[28]  Tao Qin,et al.  LETOR: A benchmark collection for research on learning to rank for information retrieval , 2010, Information Retrieval.

[29]  Quoc V. Le,et al.  Learning to Rank with Nonsmooth Cost Functions , 2006, Neural Information Processing Systems.

[30]  Hailong Sun,et al.  Recommender systems based on ranking performance optimization , 2015, Frontiers of Computer Science.

[31]  Stephen E. Robertson,et al.  GatfordCentre for Interactive Systems ResearchDepartment of Information , 1996 .

[32]  In-Chan Choi,et al.  Generalized ensemble model for document ranking in information retrieval , 2015, Comput. Sci. Inf. Syst..

[33]  Djoerd Hiemstra,et al.  A cross-benchmark comparison of 87 learning to rank methods , 2015, Inf. Process. Manag..

[34]  C. Burges,et al.  Learning to Rank Using Classification and Gradient Boosting , 2008 .

[35]  Amnon Shashua,et al.  Ranking with Large Margin Principle: Two Approaches , 2002, NIPS.

[36]  Chris Buckley,et al.  OHSUMED: an interactive retrieval evaluation and new large test collection for research , 1994, SIGIR '94.

[37]  Rong Jin,et al.  Learning to Rank by Optimizing NDCG Measure , 2009, NIPS.

[38]  Tie-Yan Liu,et al.  Listwise approach to learning to rank: theory and algorithm , 2008, ICML '08.

[39]  Stephen E. Robertson,et al.  SoftRank: optimizing non-smooth rank metrics , 2008, WSDM '08.

[40]  Jaana Kekäläinen,et al.  Cumulated gain-based evaluation of IR techniques , 2002, TOIS.

[41]  Martha Larson,et al.  xCLiMF: optimizing expected reciprocal rank for data with multiple levels of relevance , 2013, RecSys.

[42]  Hang Li Learning to Rank for Information Retrieval and Natural Language Processing , 2011, Synthesis Lectures on Human Language Technologies.

[43]  W. Bruce Croft,et al.  A language modeling approach to information retrieval , 1998, SIGIR '98.

[44]  W. Bruce Croft,et al.  Linear feature-based models for information retrieval , 2007, Information Retrieval.