Scalable Personalised Item Ranking through Parametric Density Estimation

Learning from implicit feedback is challenging because of the difficult nature of the one-class problem: we can observe only positive examples. Most conventional methods use a pairwise ranking approach and negative samplers to cope with the one-class problem. However, such methods have two main drawbacks particularly in large-scale applications; (1) the pairwise approach is severely inefficient due to the quadratic computational cost; and (2) even recent model-based samplers (e.g. IRGAN) cannot achieve practical efficiency due to the training of an extra model. In this paper, we propose a learning-to-rank approach, which achieves convergence speed comparable to the pointwise counterpart while performing similarly to the pairwise counterpart in terms of ranking effectiveness. Our approach estimates the probability densities of positive items for each user within a rich class of distributions, viz. exponential family. In our formulation, we derive a loss function and the appropriate negative sampling distribution based on maximum likelihood estimation. We also develop a practical technique for risk approximation and a regularisation scheme. We then discuss that our single-model approach is equivalent to an IRGAN variant under a certain condition. Through experiments on real-world datasets, our approach outperforms the pointwise and pairwise counterparts in terms of effectiveness and efficiency.

[1]  Xiao Xu,et al.  IPGAN: Generating Informative Item Pairs by Adversarial Sampling , 2020, IEEE Transactions on Neural Networks and Learning Systems.

[2]  Mayu Otani,et al.  Density-Ratio Based Personalised Ranking from Implicit Feedback , 2021, WWW.

[3]  Paul N. Bennett,et al.  Approximate Nearest Neighbor Negative Contrastive Learning for Dense Text Retrieval , 2020, ICLR.

[4]  Depeng Jin,et al.  Sampler Design for Bayesian Personalized Ranking by Leveraging View Data , 2018, IEEE Transactions on Knowledge and Data Engineering.

[5]  Yong Li,et al.  Simplify and Robustify Negative Sampling for Implicit Collaborative Filtering , 2020, NeurIPS.

[6]  LightGCN , 2020, Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval.

[7]  Yiqun Liu,et al.  Jointly Non-Sampling Learning for Knowledge Graph Enhanced Recommendation , 2020, SIGIR.

[8]  William L. Hamilton,et al.  Adversarial Example Games , 2020, NeurIPS.

[9]  Walid Krichene,et al.  Neural Collaborative Filtering vs. Matrix Factorization Revisited , 2020, RecSys.

[10]  Enhong Chen,et al.  Personalized Ranking with Importance Sampling , 2020, WWW.

[11]  Yixin Cao,et al.  Reinforced Negative Sampling over Knowledge Graph for Recommendation , 2020, WWW.

[12]  Zhong Ming,et al.  CoFiGAN: Collaborative filtering by generative and discriminative training for one-class recommendation , 2020, Knowl. Based Syst..

[13]  Chun Chen,et al.  Fast Adaptively Weighted Matrix Factorization for Recommendation with Implicit Feedback , 2020, AAAI.

[14]  Xiangnan He,et al.  LightGCN: Simplifying and Powering Graph Convolution Network for Recommendation , 2020, SIGIR.

[15]  Yiqun Liu,et al.  Efficient Neural Matrix Factorization without Sampling for Recommendation , 2020, ACM Trans. Inf. Syst..

[16]  Sergey I. Nikolenko,et al.  RecVAE: A New Variational Autoencoder for Top-N Recommendations with Implicit Feedback , 2019, WSDM.

[17]  Jun Guo,et al.  Wasserstein Collaborative Filtering for Item Cold-start Recommendation , 2019, ArXiv.

[18]  Xiaoyu Du,et al.  Fast Matrix Factorization With Nonuniform Weights on Missing Data , 2018, IEEE Transactions on Neural Networks and Learning Systems.

[19]  Kai Liu,et al.  Wasserstein autoencoders for collaborative filtering , 2018, Neural Computing and Applications.

[20]  Flavian Vasile,et al.  Relaxed Softmax for learning from Positive and Unlabeled data , 2019, ArXiv.

[21]  Jian Yu,et al.  Deep generative ranking for personalized recommendation , 2019, RecSys.

[22]  Depeng Jin,et al.  Reinforced Negative Sampling for Recommendation with Exposure Data , 2019, IJCAI.

[23]  Yiqun Liu,et al.  An Efficient Adaptive Transfer Neural Network for Social-aware Recommendation , 2019, SIGIR.

[24]  Haoyu Wang,et al.  Adversarial Binary Collaborative Filtering for Implicit Feedback , 2019, AAAI.

[25]  Tat-Seng Chua,et al.  Neural Graph Collaborative Filtering , 2019, SIGIR.

[26]  Yan Feng,et al.  SamWalker: Social Recommendation with Informative Sampling Strategy , 2019, WWW.

[27]  Jiliang Tang,et al.  Deep Adversarial Social Recommendation , 2019, IJCAI.

[28]  Le Song,et al.  Exponential Family Estimation via Adversarial Dynamics Embedding , 2019, NeurIPS.

[29]  Xiao Yang,et al.  WassRank: Listwise Document Ranking Using Optimal Transport Theory , 2019, WSDM.

[30]  Le Song,et al.  Kernel Exponential Family Estimation via Doubly Dual Embedding , 2018, AISTATS.

[31]  David Tse,et al.  A Convex Duality Framework for GANs , 2018, NeurIPS.

[32]  Zi Huang,et al.  Neural Memory Streaming Recommender Networks with Adversarial Training , 2018, KDD.

[33]  Matthew D. Hoffman,et al.  Variational Autoencoders for Collaborative Filtering , 2018, WWW.

[34]  Guy Blanc,et al.  Adaptive Sampled Softmax with Kernel Based Sampling , 2017, ICML.

[35]  Arthur Gretton,et al.  Kernel Conditional Exponential Family , 2017, AISTATS.

[36]  Peng Zhang,et al.  IRGAN: A Minimax Game for Unifying Generative and Discriminative Information Retrieval Models , 2017, SIGIR.

[37]  Tat-Seng Chua,et al.  Neural Collaborative Filtering , 2017, WWW.

[38]  Max Welling,et al.  Semi-Supervised Classification with Graph Convolutional Networks , 2016, ICLR.

[39]  Aapo Hyvärinen,et al.  Density Estimation in Infinite Dimensional Exponential Families , 2013, J. Mach. Learn. Res..

[40]  Jonathan M. Borwein A very complicated proof of the minimax theorem , 2016 .

[41]  Cheng Soon Ong,et al.  Linking losses for density ratio and class-probability estimation , 2016, ICML.

[42]  Julian J. McAuley,et al.  Ups and Downs: Modeling the Visual Evolution of Fashion Trends with One-Class Collaborative Filtering , 2016, WWW.

[43]  David M. Blei,et al.  Modeling User Exposure in Recommendation , 2015, WWW.

[44]  Stéphan Clémençon,et al.  Scaling-up Empirical Risk Minimization: Optimization of Incomplete $U$-statistics , 2015, J. Mach. Learn. Res..

[45]  E. Ostermann Convergence of probability measures , 2022 .

[46]  Stéphan Clémençon,et al.  SGD Algorithms based on Incomplete U-statistics: Large-Scale Minimization of Empirical Risk , 2015, NIPS.

[47]  Konstantina Christakopoulou,et al.  Collaborative Ranking with a Push at the Top , 2015, WWW.

[48]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.

[49]  Steffen Rendle,et al.  Improving pairwise learning for item recommendation from implicit feedback , 2014, WSDM.

[50]  Jason Weston,et al.  Learning to rank recommendations with the k-order statistic loss , 2013, RecSys.

[51]  Jun Wang,et al.  Optimizing top-n collaborative filtering via dynamic negative item sampling , 2013, SIGIR.

[52]  Stephen P. Boyd,et al.  Accuracy at the Top , 2012, NIPS.

[53]  Chong Wang,et al.  Latent Collaborative Retrieval , 2012, ICML.

[54]  Masashi Sugiyama,et al.  Density-ratio matching under the Bregman divergence: a unified framework of density-ratio estimation , 2012 .

[55]  Shuang-Hong Yang,et al.  Collaborative competitive filtering: learning recommender using context of user choice , 2011, SIGIR.

[56]  R. Bass,et al.  Review: P. Billingsley, Convergence of probability measures , 1971 .

[57]  Shivani Agarwal,et al.  The Infinite Push: A New Support Vector Ranking Algorithm that Directly Optimizes Accuracy at the Absolute Top of the List , 2011, SDM.

[58]  Yehuda Koren,et al.  Matrix Factorization Techniques for Recommender Systems , 2009, Computer.

[59]  Rong Pan,et al.  Mind the gaps: weighting the unknown in large-scale one-class collaborative filtering , 2009, KDD.

[60]  Lars Schmidt-Thieme,et al.  BPR: Bayesian Personalized Ranking from Implicit Feedback , 2009, UAI.

[61]  Qiang Yang,et al.  One-Class Collaborative Filtering , 2008, 2008 Eighth IEEE International Conference on Data Mining.

[62]  Yifan Hu,et al.  Collaborative Filtering for Implicit Feedback Datasets , 2008, 2008 Eighth IEEE International Conference on Data Mining.

[63]  C. Villani Optimal Transport: Old and New , 2008 .

[64]  Alexander J. Smola,et al.  Kernel methods and the exponential family , 2006, ESANN.

[65]  Gregory N. Hullender,et al.  Learning to rank using gradient descent , 2005, ICML.

[66]  Andrew W. Moore,et al.  An Investigation of Practical Approximate Nearest Neighbor Algorithms , 2004, NIPS.

[67]  Jaana Kekäläinen,et al.  Cumulated gain-based evaluation of IR techniques , 2002, TOIS.

[68]  C. Manski MAXIMUM SCORE ESTIMATION OF THE STOCHASTIC UTILITY MODEL OF CHOICE , 1975 .

[69]  D. McFadden Conditional logit analysis of qualitative choice behavior , 1972 .

[70]  R. A. Leibler,et al.  On Information and Sufficiency , 1951 .