How to Retrain Recommender System?: A Sequential Meta-Learning Method

Practical recommender systems need be periodically retrained to refresh the model with new interaction data. To pursue high model fidelity, it is usually desirable to retrain the model on both historical and new data, since it can account for both long-term and short-term user preference. However, a full model retraining could be very time-consuming and memory-costly, especially when the scale of historical data is large. In this work, we study the model retraining mechanism for recommender systems, a topic of high practical values but has been relatively little explored in the research community. Our first belief is that retraining the model on historical data is unnecessary, since the model has been trained on it before. Nevertheless, normal training on new data only may easily cause overfitting and forgetting issues, since the new data is of a smaller scale and contains fewer information on long-term user preference. To address this dilemma, we propose a new training method, aiming to abandon the historical data during retraining through learning to transfer the past training experience.Specifically, we design a neural network-based transfer component, which transforms the old model to a new model that is tailored for future recommendations. To learn the transfer component well, we optimize the "future performance'' -- i.e., the recommendation accuracy evaluated in the next time period. Our Sequential Meta-Learning(SML) method offers a general training paradigm that is applicable to any differentiable model. We demonstrate SML on matrix factorization and conduct experiments on two real-world datasets. Empirical results show that SML not only achieves significant speed-up, but also outperforms the full model retraining in recommendation accuracy, validating the effectiveness of our proposals. We release our codes at: https://github.com/zyang1580/SML.

[1]  Charu C. Aggarwal,et al.  Recommendations For Streaming Data , 2016, CIKM.

[2]  Marc'Aurelio Ranzato,et al.  Gradient Episodic Memory for Continual Learning , 2017, NIPS.

[3]  Hyunsouk Cho,et al.  MeLU: Meta-Learned User Preference Estimator for Cold-Start Recommendation , 2019, KDD.

[4]  Xiangnan He,et al.  ATM: An Attentive Translation Model for Next-Item Recommendation , 2020, IEEE Transactions on Industrial Informatics.

[5]  Julian J. McAuley,et al.  Translation-based Recommendation , 2017, RecSys.

[6]  Joemon M. Jose,et al.  A Simple Convolutional Generative Network for Next Item Recommendation , 2018, WSDM.

[7]  Dietmar Jannach,et al.  Sequence-aware Recommender Systems , 2018, UMAP.

[8]  Xing Xie,et al.  xDeepFM: Combining Explicit and Implicit Feature Interactions for Recommender Systems , 2018, KDD.

[9]  Razvan Pascanu,et al.  Overcoming catastrophic forgetting in neural networks , 2016, Proceedings of the National Academy of Sciences.

[10]  Qing He,et al.  Warm Up Cold-start Advertisements: Improving CTR Predictions via Learning to Learn ID Embeddings , 2019, SIGIR.

[11]  Ke Wang,et al.  Personalized Top-N Sequential Recommendation via Convolutional Sequence Embedding , 2018, WSDM.

[12]  Xiaoyu Du,et al.  Modeling Embedding Dimension Correlations via Convolutional Neural Collaborative Filtering , 2019, ACM Trans. Inf. Syst..

[13]  Jia Li,et al.  Latent Cross: Making Use of Context in Recurrent Recommender Systems , 2018, WSDM.

[14]  Hugo Larochelle,et al.  Optimization as a Model for Few-Shot Learning , 2016, ICLR.

[15]  Jon Atle Gulla,et al.  The Adressa dataset for news recommendation , 2017, WI.

[16]  Xiangnan He,et al.  Future Data Helps Training: Modeling Future Contexts for Session-based Recommendation , 2020, WWW.

[17]  Lars Schmidt-Thieme,et al.  BPR: Bayesian Personalized Ranking from Implicit Feedback , 2009, UAI.

[18]  Kurt Hornik,et al.  Approximation capabilities of multilayer feedforward networks , 1991, Neural Networks.

[19]  Zi Huang,et al.  Streaming Ranking Based Recommender Systems , 2018, SIGIR.

[20]  Yi Chang,et al.  Streaming Recommender Systems , 2016, WWW.

[21]  Tat-Seng Chua,et al.  Neural Factorization Machines for Sparse Predictive Analytics , 2017, SIGIR.

[22]  Zi Huang,et al.  Neural Memory Streaming Recommender Networks with Adversarial Training , 2018, KDD.

[23]  Lars Schmidt-Thieme,et al.  Online-updating regularized kernel matrix factorization models for large-scale recommender systems , 2008, RecSys '08.

[24]  Jeffrey Scott Vitter,et al.  Random sampling with a reservoir , 1985, TOMS.

[25]  Tat-Seng Chua,et al.  Neural Graph Collaborative Filtering , 2019, SIGIR.

[26]  Nicolas Kourtellis,et al.  Dynamic Matrix Factorization with Priors on Unknown Values , 2015, KDD.

[27]  Xiaowei Wang,et al.  Sequential Scenario-Specific Meta Learner for Online Recommendation , 2019, KDD.

[28]  Mubarak Shah,et al.  Task Agnostic Meta-Learning for Few-Shot Learning , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[29]  D. Sculley,et al.  Hidden Technical Debt in Machine Learning Systems , 2015, NIPS.

[30]  Yi Ren,et al.  A Survey of Personalized Recommendation Algorithm Selection Based on Meta-learning , 2019, CSIA.

[31]  Alexandros Karatzoglou,et al.  Session-based Recommendations with Recurrent Neural Networks , 2015, ICLR.

[32]  Xiangnan He,et al.  Estimation-Action-Reflection: Towards Deep Interaction Between Conversational and Recommender Systems , 2020, WSDM.

[33]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[34]  Chen Gao,et al.  λOpt: Learn to Regularize Recommender Models in Finer Levels , 2019, KDD.

[35]  Sergey Levine,et al.  Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.

[36]  Kevin Gimpel,et al.  Bridging Nonlinearities and Stochastic Regularizers with Gaussian Error Linear Units , 2016, ArXiv.

[37]  Paolo Frasconi,et al.  Bilevel Programming for Hyperparameter Optimization and Meta-Learning , 2018, ICML.

[38]  Lars Schmidt-Thieme,et al.  Real-time top-n recommendation in social streams , 2012, RecSys.

[39]  Steffen Rendle,et al.  Factorization Machines , 2010, 2010 IEEE International Conference on Data Mining.

[40]  Xiangnan He,et al.  Modeling Personalized Item Frequency Information for Next-basket Recommendation , 2020, SIGIR.

[41]  Yixin Cao,et al.  KGAT: Knowledge Graph Attention Network for Recommendation , 2019, KDD.

[42]  Tat-Seng Chua,et al.  Neural Collaborative Filtering , 2017, WWW.

[43]  Homanga Bharadhwaj,et al.  Meta-Learning for User Cold-Start Recommendation , 2019, 2019 International Joint Conference on Neural Networks (IJCNN).

[44]  Yongdong Zhang,et al.  LightGCN: Simplifying and Powering Graph Convolution Network for Recommendation , 2020, SIGIR.

[45]  Hugo Larochelle,et al.  A Meta-Learning Perspective on Cold-Start Recommendations for Items , 2017, NIPS.

[46]  Tat-Seng Chua,et al.  Fast Matrix Factorization for Online Recommendation with Implicit Feedback , 2016, SIGIR.

[47]  Lars Schmidt-Thieme,et al.  Factorizing personalized Markov chains for next-basket recommendation , 2010, WWW '10.