Top-k Consistency of Learning to Rank Methods

This paper is concerned with the consistency analysis on lis twise ranking methods. Among various ranking methods, the listwise methods ha ve competitive performances on benchmark datasets and are regarded as one of th e state-of-the-art approaches. Most listwise ranking methods manage to optimi ze ranking on the whole list (permutation) of objects, however, in practical applications such as information retrieval, correct ranking at the top k positions is much more important. This paper aims to analyze whether existing listwise rankin g methods are statistically consistent in the topk setting. For this purpose, we define a topk ranking framework, where the true loss (and thus the risks) are define d o the basis of topk subgroup of permutations. This framework can include the pe rmutation-level ranking framework proposed in previous work as a special cas e. Based on the new framework, we derive sufficient conditions for a listwis e ranking method to be consistent with the topk true loss, and show an effective way of modifying the surrogate loss functions in existing methods to satisfy the se conditions. Experimental results show that after the modification, the methods can work significantly better than their original versions, indicating the correc tn ss of our theoretical analysis.

[1]  Tie-Yan Liu,et al.  Listwise approach to learning to rank: theory and algorithm , 2008, ICML '08.

[2]  Tie-Yan Liu,et al.  Learning to rank: from pairwise approach to listwise approach , 2007, ICML '07.

[3]  Yoram Singer,et al.  An Efficient Boosting Algorithm for Combining Preferences by , 2013 .

[4]  Cynthia Rudin,et al.  Ranking with a P-Norm Push , 2006, COLT.

[5]  Tong Zhang,et al.  Statistical Analysis of Some Multi-Category Large Margin Classification Methods , 2004, J. Mach. Learn. Res..

[6]  Gregory N. Hullender,et al.  Learning to rank using gradient descent , 2005, ICML.

[7]  Tao Qin,et al.  Query-level loss functions for information retrieval , 2008, Inf. Process. Manag..

[8]  J. Marden Analyzing and Modeling Rank Data , 1996 .

[9]  Qiang Wu,et al.  McRank: Learning to Rank Using Multiple Classification and Gradient Boosting , 2007, NIPS.

[10]  Klaus Obermayer,et al.  Support vector learning for ordinal regression , 1999 .

[11]  Tong Zhang,et al.  Statistical Analysis of Bayes Optimal Subset Ranking , 2008, IEEE Transactions on Information Theory.

[12]  Tao Qin,et al.  LETOR: Benchmark Dataset for Research on Learning to Rank for Information Retrieval , 2007 .

[13]  Ramesh Nallapati,et al.  Discriminative models for information retrieval , 2004, SIGIR '04.

[14]  Michael I. Jordan,et al.  Convexity, Classification, and Risk Bounds , 2006 .

[15]  Tong Zhang,et al.  Subset Ranking Using Regression , 2006, COLT.

[16]  Stéphan Clémençon,et al.  Ranking the Best Instances , 2006, J. Mach. Learn. Res..