Privileged Graph Distillation for Cold Start Recommendation

The cold start problem in recommender systems is a long-standing challenge, which requires recommending to new users (items) based on attributes without any historical interaction records. In these recommendation systems, warm users (items) have privileged collaborative signals of interaction records compared to cold start users (items), and these Collaborative Filtering (CF) signals are shown to have competing performance for recommendation. Many researchers proposed to learn the correlation between collaborative signal embedding space and the attribute embedding space to improve the cold start recommendation, in which user and item categorical attributes are available in many online platforms. However, the cold start recommendation is still limited by two embedding spaces modeling and simple assumptions of space transformation. As user-item interaction behaviors and user (item) attributes naturally form a heterogeneous graph structure, in this paper, we propose a privileged graph distillation model (PGD). The teacher model is composed of a heterogeneous graph structure for warm users and items with privileged CF links. The student model is composed of an entity-attribute graph without CF links. Specifically, the teacher model can learn better embeddings of each entity by injecting complex higher-order relationships from the constructed heterogeneous graph. The student model can learn the distilled output with privileged CF embeddings from the teacher embeddings. Our proposed model is generally applicable to different cold start scenarios with new user, new item, or new user-new item. Finally, extensive experimental results on the real-world datasets clearly show the effectiveness of our proposed model on different types of cold start problems, with average 6.6%, 5.6%, and 17.1% improvement over state-of-the-art baselines on three datasets, respectively.

[1]  Ke Wang,et al.  Ranking Distillation: Learning Compact Ranking Models With High Performance for Recommender System , 2018, KDD.

[2]  Le Wu,et al.  A Neural Influence Diffusion Model for Social Recommendation , 2019, SIGIR.

[3]  Meng Wang,et al.  Revisiting Graph based Collaborative Filtering: A Linear Residual Graph Convolutional Network Approach , 2020, AAAI.

[4]  Xing Xie,et al.  xDeepFM: Combining Explicit and Implicit Feature Interactions for Recommender Systems , 2018, KDD.

[5]  Fabian Abel,et al.  RecSys Challenge 2017: Offline and Online Evaluation , 2017, RecSys.

[6]  Dit-Yan Yeung,et al.  Collaborative Deep Learning for Recommender Systems , 2014, KDD.

[7]  Geoffrey E. Hinton,et al.  Distilling the Knowledge in a Neural Network , 2015, ArXiv.

[8]  Jure Leskovec,et al.  Hyperbolic Graph Convolutional Neural Networks , 2019, NeurIPS.

[9]  Rajgopal Kannan,et al.  GraphSAINT: Graph Sampling Based Inductive Learning Method , 2019, ICLR.

[10]  Jure Leskovec,et al.  Inductive Representation Learning on Large Graphs , 2017, NIPS.

[11]  Yan Lu,et al.  Relational Knowledge Distillation , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[12]  Zi Huang,et al.  Next Point-of-Interest Recommendation on Resource-Constrained Mobile Devices , 2020, WWW.

[13]  Max Welling,et al.  Graph Convolutional Matrix Completion , 2017, ArXiv.

[14]  Lars Schmidt-Thieme,et al.  Learning Attribute-to-Feature Mappings for Cold-Start Recommendations , 2010, 2010 IEEE International Conference on Data Mining.

[15]  Seyed Iman Mirzadeh,et al.  Improved Knowledge Distillation via Teacher Assistant , 2020, AAAI.

[16]  Tat-Seng Chua,et al.  Neural Graph Collaborative Filtering , 2019, SIGIR.

[17]  Bowei Chen,et al.  Multi-view Visual Bayesian Personalized Ranking from Implicit Feedback , 2018, UMAP.

[18]  M. de Rijke,et al.  Long Short-Term Session Search: Joint Personalized Reranking and Next Query Prediction , 2021, WWW.

[19]  Shuang-Hong Yang,et al.  Functional matrix factorizations for cold-start recommendation , 2011, SIGIR.

[20]  Julian J. McAuley,et al.  VBPR: Visual Bayesian Personalized Ranking from Implicit Feedback , 2015, AAAI.

[21]  Yixin Chen,et al.  Inductive Graph Pattern Learning for Recommender Systems Based on a Graph Neural Network , 2019, ArXiv.

[22]  Yongdong Zhang,et al.  LightGCN: Simplifying and Powering Graph Convolution Network for Recommendation , 2020, SIGIR.

[23]  Bing Li,et al.  Knowledge Distillation via Instance Relationship Graph , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[24]  Benjamin Schrauwen,et al.  Deep content-based music recommendation , 2013, NIPS.

[25]  Scott Sanner,et al.  Social collaborative filtering for cold-start recommendations , 2014, RecSys '14.

[26]  Hanning Zhou,et al.  Distilling Structured Knowledge into Embeddings for Explainable and Accurate Recommendation , 2019, WSDM.

[27]  M. de Rijke,et al.  Social Collaborative Viewpoint Regression with Explainable Recommendations , 2017, WSDM.

[28]  Xu Chen,et al.  Adversarial Distillation for Efficient Recommendation with External Knowledge , 2018, ACM Trans. Inf. Syst..

[29]  Yoshua Bengio,et al.  FitNets: Hints for Thin Deep Nets , 2014, ICLR.

[30]  Jure Leskovec,et al.  Graph Convolutional Neural Networks for Web-Scale Recommender Systems , 2018, KDD.

[31]  Ruslan Salakhutdinov,et al.  Bayesian probabilistic matrix factorization using Markov chain Monte Carlo , 2008, ICML '08.

[32]  Gang Fu,et al.  Deep & Cross Network for Ad Click Predictions , 2017, ADKDD@KDD.

[33]  Kenneth Y. Goldberg,et al.  Eigentaste: A Constant Time Collaborative Filtering Algorithm , 2001, Information Retrieval.

[34]  Lei Chen,et al.  Learning to Transfer Graph Embeddings for Inductive Graph based Recommendation , 2020, SIGIR.

[35]  Guorui Zhou,et al.  Rocket Launching: A Universal and Efficient Framework for Training Well-performing Light Net , 2017, AAAI.

[36]  Hwanjo Yu,et al.  DE-RRD: A Knowledge Distillation Framework for Recommender System , 2020, CIKM.

[37]  Dacheng Tao,et al.  Learning Student Networks via Feature Embedding , 2018, IEEE Transactions on Neural Networks and Learning Systems.

[38]  Maksims Volkovs,et al.  DropoutNet: Addressing Cold Start in Recommender Systems , 2017, NIPS.

[39]  Yehuda Koren,et al.  Matrix Factorization Techniques for Recommender Systems , 2009, Computer.

[40]  Jian Wu,et al.  Privileged Features Distillation at Taobao Recommendations , 2020, KDD.

[41]  Yu Liu,et al.  Correlation Congruence for Knowledge Distillation , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).

[42]  Nikos Komodakis,et al.  Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer , 2016, ICLR.

[43]  Bernhard Schölkopf,et al.  Unifying distillation and privileged information , 2015, ICLR.

[44]  James Caverlee,et al.  Recommendation for New Users and New Items via Randomized Training and Mixture-of-Experts Transformation , 2020, SIGIR.