Diversification-Aware Learning to Rank using Distributed Representation

Existing work on search result diversification typically falls into the “next document” paradigm, that is, selecting the next document based on the ones already chosen. A sequential process of selecting documents one-by-one is naturally modeled in learning-based approaches. However, such a process makes the learning difficult because there are an exponential number of ranking lists to consider. Sampling is usually used to reduce the computational complexity but this makes the learning less effective. In this paper, we propose a soft version of the “next document” paradigm in which we associate each document with an approximate rank, and thus the subtopics covered prior to a document can also be estimated. We show that we can derive differentiable diversification-aware losses, which are smooth approximation of diversity metrics like α-NDCG, based on these estimates. We further propose to optimize the losses in the learning-to-rank setting using neural distributed representations of queries and documents. Experiments are conducted on the public benchmark TREC datasets. By comparing with an extensive list of baseline methods, we show that our Diversification-Aware LEarning-TO-Rank (DALETOR) approaches outperform them by a large margin, while being much simpler during learning and inference.

[1]  Shuguang Han,et al.  A Stochastic Treatment of Learning to Rank Scoring Functions , 2020, WSDM.

[2]  Sebastian Bruch,et al.  Learning Groupwise Multivariate Scoring Functions Using Deep Neural Networks , 2018, ICTIR.

[3]  Yi Tay,et al.  Are Neural Rankers still Outperformed by Gradient Boosted Decision Trees? , 2021, ICLR.

[4]  Andrei Z. Broder,et al.  Identifying and Filtering Near-Duplicate Documents , 2000, CPM.

[5]  Quoc V. Le,et al.  Distributed Representations of Sentences and Documents , 2014, ICML.

[6]  W. Bruce Croft,et al.  Diversity by proportionality: an election-based approach to search result diversification , 2012, SIGIR '12.

[7]  Özgür Ulusoy,et al.  Supervised approaches for explicit search result diversification , 2020, Inf. Process. Manag..

[8]  Thorsten Joachims,et al.  Predicting diverse subsets using structural SVMs , 2008, ICML '08.

[9]  Wei Zeng,et al.  Adapting Markov Decision Process for Search Result Diversification , 2017, SIGIR.

[10]  Larry P. Heck,et al.  Learning deep structured semantic models for web search using clickthrough data , 2013, CIKM.

[11]  Xueqi Cheng,et al.  Learning Maximal Marginal Relevance Model via Directly Optimizing Diversity Evaluation Measures , 2015, SIGIR.

[12]  Jade Goldstein-Stewart,et al.  The use of MMR, diversity-based reranking for reordering documents and producing summaries , 1998, SIGIR '98.

[13]  Sebastian Bruch,et al.  TF-Ranking: Scalable TensorFlow Library for Learning-to-Rank , 2018, KDD.

[14]  Xueqi Cheng,et al.  Modeling Document Novelty with Neural Tensor Network for Search Result Diversification , 2016, SIGIR.

[15]  Yoram Singer,et al.  Adaptive Subgradient Methods for Online Learning and Stochastic Optimization , 2011, J. Mach. Learn. Res..

[16]  Jun Xu,et al.  SetRank: Learning a Permutation-Invariant Ranking Model for Information Retrieval , 2020, SIGIR.

[17]  Xuanhui Wang,et al.  Learning-to-Rank with BERT in TF-Ranking , 2020, ArXiv.

[18]  Michael Bendersky,et al.  Permutation Equivariant Document Interaction Network for Neural Learning to Rank , 2020, ICTIR.

[19]  Ji-Rong Wen,et al.  Diversifying Search Results using Self-Attention Network , 2020, CIKM.

[20]  Olivier Chapelle,et al.  Expected reciprocal rank for graded relevance , 2009, CIKM.

[21]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[22]  Ji-Rong Wen,et al.  Learning to Diversify Search Results via Subtopic Attention , 2017, SIGIR.

[23]  Jia Li,et al.  Latent Cross: Making Use of Context in Recurrent Recommender Systems , 2018, WSDM.

[24]  Wei Zeng,et al.  From Greedy Selection to Exploratory Decision-Making: Diverse Ranking with Policy-Value Networks , 2018, SIGIR.

[25]  Ajith Ramanathan,et al.  Practical Diversified Recommendations on YouTube with Determinantal Point Processes , 2018, CIKM.

[26]  Stefano Ermon,et al.  Stochastic Optimization of Sorting Networks via Continuous Relaxations , 2019, ICLR.

[27]  Tie-Yan Liu,et al.  LightGBM: A Highly Efficient Gradient Boosting Decision Tree , 2017, NIPS.

[28]  W. Bruce Croft,et al.  Learning a Deep Listwise Context Model for Ranking Refinement , 2018, SIGIR.

[29]  Charles L. A. Clarke,et al.  Novelty and diversity in information retrieval evaluation , 2008, SIGIR '08.

[30]  Filip Radlinski,et al.  Learning optimally diverse rankings over large document collections , 2010, ICML.

[31]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[32]  Xueqi Cheng,et al.  Learning for search result diversification , 2014, SIGIR.

[33]  Tie-Yan Liu,et al.  Learning to rank for information retrieval , 2009, SIGIR.

[34]  Pinar Donmez,et al.  On the local optimality of LambdaRank , 2009, SIGIR.

[35]  Sreenivas Gollapudi,et al.  Diversifying search results , 2009, WSDM '09.

[36]  Ming-Wei Chang,et al.  BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.

[37]  Craig MacDonald,et al.  Exploiting query reformulations for web search result diversification , 2010, WWW '10.

[38]  Ji-Rong Wen,et al.  DVGAN: A Minimax Game for Search Result Diversification Combining Explicit and Implicit Features , 2020, SIGIR.

[39]  Tao Qin,et al.  A general approximation framework for direct optimization of information retrieval measures , 2010, Information Retrieval.

[40]  Cheng Li,et al.  The LambdaLoss Framework for Ranking Metric Optimization , 2018, CIKM.