Contrastive Self-supervised Sequential Recommendation with Robust Augmentation

Sequential Recommendation describes a set of techniques to model dynamic user behavior in order to predict future interactions in sequential user data. At their core, such approaches model transition probabilities between items in a sequence, whether through Markov chains, recurrent networks, or more recently, Transformers. However both old and new issues remain, including data-sparsity and noisy data; such issues can impair performance, especially in complex, parameter-hungry models. In this paper, we investigate the application of contrastive Self-Supervised Learning (SSL) to sequential recommendation, as a way to alleviate some of these issues. Contrastive SSL constructs augmentations from unlabelled instances, where agreements among positive pairs are maximized. It is challenging to devise a contrastive SSL framework for sequential recommendation, due to its discrete nature, correlations among items, and skewness of length distributions. To this end, we propose a novel framework, Contrastive Self-supervised Learning for Sequential Recommendation (CoSeRec). We introduce two informative augmentation operators leveraging item correlations to create high quality views for contrastive learning. Experimental results on three real-world datasets demonstrate the effectiveness of the proposed method on improving model performance, and the robustness against sparse and noisy data. Our implementation is available: https://github.com/YChen1993/CoSeRec

[1]  Kaiming He,et al.  Momentum Contrast for Unsupervised Visual Representation Learning , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[2]  Lars Schmidt-Thieme,et al.  BPR: Bayesian Personalized Ranking from Implicit Feedback , 2009, UAI.

[3]  Zhiwei Liu,et al.  Modeling Sequences as Distributions with Uncertainty for Sequential Recommendation , 2021, CIKM.

[4]  Feng Yu,et al.  A Dynamic Recurrent Model for Next Basket Recommendation , 2016, SIGIR.

[5]  Alexei A. Efros,et al.  What Should Not Be Contrastive in Contrastive Learning , 2020, ICLR.

[6]  Gary D. Bader,et al.  DeCLUTR: Deep Contrastive Learning for Unsupervised Textual Representations , 2020, ACL.

[7]  Walid Krichene,et al.  On Sampled Metrics for Item Recommendation , 2020, KDD.

[8]  Philip S. Yu,et al.  Augmenting Sequential Recommendation with Pseudo-Prior Items via Reversely Pre-training Transformer , 2021, SIGIR.

[9]  Yujie Wang,et al.  Time Interval Aware Self-Attention for Sequential Recommendation , 2020, WSDM.

[10]  Pietro Liò,et al.  Deep Graph Infomax , 2018, ICLR.

[11]  Madian Khabsa,et al.  CLEAR: Contrastive Learning for Sentence Representation , 2020, ArXiv.

[12]  Philip S. Yu,et al.  JSCN: Joint Spectral Convolutional Network for Cross Domain Recommendation , 2019, 2019 IEEE International Conference on Big Data (Big Data).

[13]  Kai Zou,et al.  EDA: Easy Data Augmentation Techniques for Boosting Performance on Text Classification Tasks , 2019, EMNLP.

[14]  Kevin Gimpel,et al.  ALBERT: A Lite BERT for Self-supervised Learning of Language Representations , 2019, ICLR.

[15]  Cho-Jui Hsieh,et al.  SSE-PT: Sequential Recommendation Via Personalized Transformer , 2020, RecSys.

[16]  Anton van den Hengel,et al.  Image-Based Recommendations on Styles and Substitutes , 2015, SIGIR.

[17]  Chang Zhou,et al.  Contrastive Learning for Debiased Candidate Generation at Scale , 2020, ArXiv.

[18]  Jieqi Kang,et al.  Self-supervised Learning for Deep Models in Recommendations , 2020, ArXiv.

[19]  Zi Huang,et al.  Try This Instead: Personalized and Interpretable Substitute Recommendation , 2020, SIGIR.

[20]  Peng Jiang,et al.  BERT4Rec: Sequential Recommendation with Bidirectional Encoder Representations from Transformer , 2019, CIKM.

[21]  Julian J. McAuley,et al.  Fusing Similarity Models with Markov Chains for Sparse Sequential Recommendation , 2016, 2016 IEEE 16th International Conference on Data Mining (ICDM).

[22]  Jiawei Zhang,et al.  Continuous-Time Sequential Recommendation with Temporal Graph Collaborative Transformer , 2021, CIKM.

[23]  Yongfeng Zhang,et al.  Sequential Recommendation with User Memory Networks , 2018, WSDM.

[24]  Lars Schmidt-Thieme,et al.  Factorizing personalized Markov chains for next-basket recommendation , 2010, WWW '10.

[25]  Hongning Wang,et al.  Déjà vu: A Contextualized Temporal Attention Mechanism for Sequential Recommendation , 2020, WWW.

[26]  Zheng Qin,et al.  Time Matters: Sequential Recommendation with Complex Temporal Information , 2020, SIGIR.

[27]  Alexandros Karatzoglou,et al.  Personalizing Session-based Recommendations with Hierarchical Recurrent Neural Networks , 2017, RecSys.

[28]  Tat-Seng Chua,et al.  Denoising Implicit Feedback for Recommendation , 2020, WSDM.

[29]  Qiang Liu,et al.  Graph Contrastive Learning with Adaptive Augmentation , 2020, WWW.

[30]  Pengtao Xie,et al.  CERT: Contrastive Self-supervised Learning for Language Understanding , 2020, ArXiv.

[31]  Ming-Wei Chang,et al.  BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.

[32]  Bolin Ding,et al.  Contrastive Pre-training for Sequential Recommendation , 2020, ArXiv.

[33]  Yizhu Jiao,et al.  Sub-graph Contrast for Scalable Self-Supervised Graph Representation Learning , 2020, 2020 IEEE International Conference on Data Mining (ICDM).

[34]  David Heckerman,et al.  Empirical Analysis of Predictive Algorithms for Collaborative Filtering , 1998, UAI.

[35]  Jianxun Lian,et al.  Self-supervised Graph Learning for Recommendation , 2020, SIGIR.

[36]  Cordelia Schmid,et al.  What makes for good views for contrastive learning , 2020, NeurIPS.

[37]  Philip S. Yu,et al.  BasConv: Aggregating Heterogeneous Interactions for Basket Recommendation with Graph Convolutional Neural Network , 2020, SDM.

[38]  Yuxiao Dong,et al.  GCC: Graph Contrastive Coding for Graph Neural Network Pre-Training , 2020, KDD.

[39]  Aapo Hyvärinen,et al.  Noise-contrastive estimation: A new estimation principle for unnormalized statistical models , 2010, AISTATS.

[40]  Julian J. McAuley,et al.  Self-Attentive Sequential Recommendation , 2018, 2018 IEEE International Conference on Data Mining (ICDM).

[41]  Tat-Seng Chua,et al.  Neural Graph Collaborative Filtering , 2019, SIGIR.

[42]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[43]  Chang Zhou,et al.  Disentangled Self-Supervision in Sequential Recommenders , 2020, KDD.

[44]  Ke Wang,et al.  Personalized Top-N Sequential Recommendation via Convolutional Sequence Embedding , 2018, WSDM.

[45]  Tianyu Gao,et al.  SimCSE: Simple Contrastive Learning of Sentence Embeddings , 2021, EMNLP.

[46]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[47]  Ji-Rong Wen,et al.  S3-Rec: Self-Supervised Learning for Sequential Recommendation with Mutual Information Maximization , 2020, CIKM.

[48]  Laurens van der Maaten,et al.  Self-Supervised Learning of Pretext-Invariant Representations , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[49]  Zhangyang Wang,et al.  Graph Contrastive Learning with Augmentations , 2020, NeurIPS.

[50]  Geoffrey E. Hinton,et al.  A Simple Framework for Contrastive Learning of Visual Representations , 2020, ICML.

[51]  Alex Beutel,et al.  Recurrent Recommender Networks , 2017, WSDM.

[52]  Alexandros Karatzoglou,et al.  Session-based Recommendations with Recurrent Neural Networks , 2015, ICLR.