A Domain Generalization Perspective on Listwise Context Modeling

As one of the most popular techniques for solving the ranking problem in information retrieval, Learning-to-rank (LETOR) has received a lot of attention both in academia and industry due to its importance in a wide variety of data mining applications. However, most of existing LETOR approaches choose to learn a single global ranking function to handle all queries, and ignore the substantial differences that exist between queries. In this paper, we propose a domain generalization strategy to tackle this problem. We propose Query-Invariant Listwise Context Modeling (QILCM), a novel neural architecture which eliminates the detrimental influence of inter-query variability by learning \textit{query-invariant} latent representations, such that the ranking system could generalize better to unseen queries. We evaluate our techniques on benchmark datasets, demonstrating that QILCM outperforms previous state-of-the-art approaches by a substantial margin.

[1]  Stephen E. Robertson,et al.  SoftRank: optimizing non-smooth rank metrics , 2008, WSDM '08.

[2]  Bowen Zhou,et al.  A Structured Self-attentive Sentence Embedding , 2017, ICLR.

[3]  Shubhra Kanti Karmaker Santu,et al.  On Application of Learning to Rank for E-Commerce Search , 2017, SIGIR.

[4]  Ramesh Nallapati,et al.  Discriminative models for information retrieval , 2004, SIGIR '04.

[5]  Bhaskar Mitra,et al.  Cross Domain Regularization for Neural Ranking Models using Adversarial Learning , 2018, SIGIR.

[6]  Qiang Wu,et al.  McRank: Learning to Rank Using Multiple Classification and Gradient Boosting , 2007, NIPS.

[7]  Antonino Freno,et al.  Practical Lessons from Developing a Large-Scale Recommender System at Zalando , 2017, RecSys.

[8]  Donald A. Adjeroh,et al.  Unified Deep Supervised Domain Adaptation and Generalization , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[9]  Fabrizio Silvestri,et al.  Post-Learning Optimization of Tree Ensembles for Efficient Ranking , 2016, SIGIR.

[10]  Bhaskar Mitra,et al.  Neural Ranking Models with Multiple Document Fields , 2018, WSDM.

[11]  Jun Xu,et al.  Modeling Diverse Relevance Patterns in Ad-hoc Retrieval , 2018, SIGIR.

[12]  Raffaele Perego,et al.  Quality versus efficiency in document scoring with learning-to-rank models , 2016, Inf. Process. Manag..

[13]  Sepp Hochreiter,et al.  Fast and Accurate Deep Network Learning by Exponential Linear Units (ELUs) , 2015, ICLR.

[14]  Alejandro Mottini,et al.  Deep Choice Model Using Pointer Networks for Airline Itinerary Prediction , 2017, KDD.

[15]  Zhi-Hua Zhou,et al.  Isolation Forest , 2008, 2008 Eighth IEEE International Conference on Data Mining.

[16]  W. Bruce Croft,et al.  Learning a Deep Listwise Context Model for Ranking Refinement , 2018, SIGIR.

[17]  Sergey Ioffe,et al.  Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.

[18]  Siddhartha Chaudhuri,et al.  Generalizing Across Domains via Cross-Gradient Training , 2018, ICLR.

[19]  Gregory N. Hullender,et al.  Learning to rank using gradient descent , 2005, ICML.

[20]  François Laviolette,et al.  Domain-Adversarial Training of Neural Networks , 2015, J. Mach. Learn. Res..

[21]  Cho-Jui Hsieh,et al.  SQL-Rank: A Listwise Approach to Collaborative Ranking , 2018, ICML.

[22]  Paul Covington,et al.  Deep Neural Networks for YouTube Recommendations , 2016, RecSys.

[23]  Tao Qin,et al.  Introducing LETOR 4.0 Datasets , 2013, ArXiv.

[24]  Hao Su,et al.  A Point Set Generation Network for 3D Object Reconstruction from a Single Image , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[25]  Craig MacDonald,et al.  A Contextual Attention Recurrent Architecture for Context-Aware Venue Recommendation , 2018, SIGIR.

[26]  Bernhard Schölkopf,et al.  Domain Generalization via Invariant Feature Representation , 2013, ICML.

[27]  Yongxin Yang,et al.  Learning to Generalize: Meta-Learning for Domain Generalization , 2017, AAAI.

[28]  Gilles Blanchard,et al.  Generalizing from Several Related Classification Tasks to a New Unlabeled Sample , 2011, NIPS.

[29]  Qiang Wu,et al.  Adapting boosting for information retrieval measures , 2010, Information Retrieval.

[30]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[32]  Harry Shum,et al.  Query Dependent Ranking Using K-nearest Neighbor * , 2022 .

[33]  Enhong Chen,et al.  Context-aware ranking in web search , 2010, SIGIR '10.

[34]  Graham Neubig,et al.  Controllable Invariance through Adversarial Feature Learning , 2017, NIPS.

[35]  Danna Zhou,et al.  d. , 1934, Microbial pathogenesis.

[36]  Maksims Volkovs,et al.  Collaborative Ranking With 17 Parameters , 2012, NIPS.

[37]  Daniel P. W. Ellis,et al.  Feed-Forward Networks with Attention Can Solve Some Long-Term Memory Problems , 2015, ArXiv.

[38]  Yongfeng Zhang,et al.  Sequential Recommendation with User Memory Networks , 2018, WSDM.

[39]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[40]  Hamed Zamani,et al.  Situational Context for Ranking in Personal Search , 2017, WWW.

[41]  Chang Zhou,et al.  ATRank: An Attention-Based User Behavior Modeling Framework for Recommendation , 2017, AAAI.

[42]  Yi Chang,et al.  Yahoo! Learning to Rank Challenge Overview , 2010, Yahoo! Learning to Rank Challenge.

[43]  Wenwu Ou,et al.  Globally Optimized Mutual Influence Aware Ranking in E-Commerce Search , 2018, IJCAI.

[44]  Tie-Yan Liu,et al.  Learning to rank: from pairwise approach to listwise approach , 2007, ICML '07.

[45]  Klaus-Robert Müller,et al.  Efficient BackProp , 2012, Neural Networks: Tricks of the Trade.

[46]  Yoram Singer,et al.  An Efficient Boosting Algorithm for Combining Preferences by , 2013 .

[47]  Yann Dauphin,et al.  Convolutional Sequence to Sequence Learning , 2017, ICML.

[48]  Samy Bengio,et al.  Order Matters: Sequence to sequence for sets , 2015, ICLR.