Listwise Learning to Rank by Exploring Unique Ratings

In this paper, we propose new listwise learning-to-rank models that mitigate the shortcomings of existing ones. Existing listwise learning-to-rank models are generally derived from the classical Plackett-Luce model, which has three major limitations. (1) Its permutation probabilities overlook ties, i.e., a situation when more than one document has the same rating with respect to a query. This can lead to imprecise permutation probabilities and inefficient training because of selecting documents one by one. (2) It does not favor documents having high relevance. (3) It has a loose assumption that sampling documents at different steps is independent. To overcome the first two limitations, we model ranking as selecting documents from a candidate set based on unique rating levels in decreasing order. The number of steps in training is determined by the number of unique rating levels. More specifically, in each step, we apply multiple multi-class classification tasks to a document candidate set and choose all documents that have the highest rating from the document set. This is in contrast to taking one document step by step in the classical Plackett-Luce model. Afterward, we remove all of the selected documents from the document set and repeat until the remaining documents all have the lowest rating. We propose a new loss function and associated four models for the entire sequence of weighted classification tasks by assigning high weights to the selected documents with high ratings for optimizing Normalized Discounted Cumulative Gain (NDCG). To overcome the final limitation, we further propose a novel and efficient way of refining prediction scores by combining an adapted Vanilla Recurrent Neural Network (RNN) model with pooling given selected documents at previous steps. We encode all of the documents already selected by an RNN model. In a single step, we rank all of the documents with the same ratings using the last cell of the RNN multiple times. We have implemented our models using three settings: neural networks, neural networks with gradient boosting, and regression trees with gradient boosting. We have conducted experiments on four public datasets. The experiments demonstrate that the models notably outperform state-of-the-art learning-to-rank models.

[1]  Xueqi Cheng,et al.  Position-Aware ListMLE: A Sequential Learning Process for Ranking , 2014, UAI.

[2]  Tie-Yan Liu,et al.  LightGBM: A Highly Efficient Gradient Boosting Decision Tree , 2017, NIPS.

[3]  Tianqi Chen,et al.  XGBoost: A Scalable Tree Boosting System , 2016, KDD.

[4]  Xueqi Cheng,et al.  A new probabilistic model for top-k ranking problem , 2012, CIKM.

[5]  M. de Rijke,et al.  Modeling Label Ambiguity for Neural List-Wise Learning to Rank , 2017, ArXiv.

[6]  Xuanhui Wang,et al.  Combining Decision Trees and Neural Networks for Learning-to-Rank in Personal Search , 2019, KDD.

[7]  Sebastian Bruch,et al.  Revisiting Approximate Metric Optimization in the Age of Deep Neural Networks , 2019, SIGIR.

[8]  Jürgen Schmidhuber,et al.  Long Short-Term Memory , 1997, Neural Computation.

[9]  David Barber,et al.  Dealing with a large number of classes -- Likelihood, Discrimination or Ranking? , 2016, 1606.06959.

[10]  W. Bruce Croft,et al.  Learning a Deep Listwise Context Model for Ranking Refinement , 2018, SIGIR.

[11]  D. Sculley,et al.  Large Scale Learning to Rank , 2009 .

[12]  Tie-Yan Liu,et al.  Listwise approach to learning to rank: theory and algorithm , 2008, ICML '08.

[13]  Stephen E. Robertson,et al.  SoftRank: optimizing non-smooth rank metrics , 2008, WSDM '08.

[14]  Jaana Kekäläinen,et al.  Cumulated gain-based evaluation of IR techniques , 2002, TOIS.

[15]  Yoshua Bengio,et al.  Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling , 2014, ArXiv.

[16]  Jiyun Luo,et al.  Dynamic Search Models and Applications , 2018 .

[17]  Tie-Yan Liu,et al.  Learning to Rank for Information Retrieval , 2011 .

[18]  R. Duncan Luce,et al.  Individual Choice Behavior: A Theoretical Analysis , 1979 .

[19]  Jiafeng Guo,et al.  Reinforcement Learning to Rank with Markov Decision Process , 2017, SIGIR.

[20]  Cheng Li,et al.  The LambdaLoss Framework for Ranking Metric Optimization , 2018, CIKM.

[21]  Adriano Veloso,et al.  Learning to Rank with Deep Autoencoder Features , 2018, 2018 International Joint Conference on Neural Networks (IJCNN).

[22]  Marc Najork,et al.  Learning Groupwise Scoring Functions Using Deep Neural Networks , 2018, ArXiv.

[23]  R. Plackett The Analysis of Permutations , 1975 .

[24]  J. Friedman Greedy function approximation: A gradient boosting machine. , 2001 .

[25]  Tao Qin,et al.  LETOR: A benchmark collection for research on learning to rank for information retrieval , 2010, Information Retrieval.

[26]  Tie-Yan Liu,et al.  Learning to rank: from pairwise approach to listwise approach , 2007, ICML '07.

[27]  Elad Eban,et al.  Seq2Slate: Re-ranking and Slate Optimization with RNNs , 2018, ArXiv.

[28]  Wei Zeng,et al.  Multi Page Search with Reinforcement Learning to Rank , 2018, ICTIR.

[29]  Tao Xu,et al.  Applying Deep Learning to Airbnb Search , 2018, KDD.

[30]  Xueqi Cheng,et al.  DeepRank: A New Deep Architecture for Relevance Ranking in Information Retrieval , 2017, CIKM.

[31]  Thore Graepel,et al.  Large Margin Rank Boundaries for Ordinal Regression , 2000 .

[32]  Yoram Singer,et al.  An Efficient Boosting Algorithm for Combining Preferences by , 2013 .

[33]  Gregory N. Hullender,et al.  Learning to rank using gradient descent , 2005, ICML.

[34]  John Guiver,et al.  Bayesian inference for Plackett-Luce ranking models , 2009, ICML '09.

[35]  Tao Qin,et al.  Introducing LETOR 4.0 Datasets , 2013, ArXiv.

[36]  M. de Rijke,et al.  Ranking for Relevance and Display Preferences in Complex Presentation Layouts , 2018, SIGIR.

[37]  Timothy A. Mann,et al.  Beyond Greedy Ranking: Slate Optimization via List-CVAE , 2018, ICLR.

[38]  Sebastian Bruch,et al.  Learning Groupwise Multivariate Scoring Functions Using Deep Neural Networks , 2018, ICTIR.

[39]  Yang Wang,et al.  Unbiased LambdaMART: An Unbiased Pairwise Learning-to-Rank Algorithm , 2018, WWW.

[40]  J. Shane Culpepper,et al.  Joint Optimization of Cascade Ranking Models , 2019, WSDM.

[41]  Feng Liu,et al.  Novel Approaches to Accelerating the Convergence Rate of Markov Decision Process for Search Result Diversification , 2018, DASFAA.

[42]  Quoc V. Le,et al.  Learning to Rank with Nonsmooth Cost Functions , 2006, Neural Information Processing Systems.

[43]  Yong Yu,et al.  Learning to rank with ties , 2008, SIGIR '08.

[44]  Ling Li,et al.  Ordinal Regression by Extended Binary Classification , 2006, NIPS.

[45]  Sebastian Bruch,et al.  TF-Ranking: Scalable TensorFlow Library for Learning-to-Rank , 2018, KDD.

[46]  Tie-Yan Liu,et al.  A Communication-Efficient Parallel Algorithm for Decision Tree , 2016, NIPS.

[47]  Olivier Chapelle,et al.  Expected reciprocal rank for graded relevance , 2009, CIKM.

[48]  Wei Zeng,et al.  From Greedy Selection to Exploratory Decision-Making: Diverse Ranking with Policy-Value Networks , 2018, SIGIR.

[49]  Christopher J. C. Burges,et al.  From RankNet to LambdaRank to LambdaMART: An Overview , 2010 .

[50]  Dong Wang,et al.  Stochastic Top-k ListNet , 2015, EMNLP.

[51]  W. Bruce Croft,et al.  A Deep Look into Neural Ranking Models for Information Retrieval , 2019, Inf. Process. Manag..