Learning Graph Meta Embeddings for Cold-Start Ads in Click-Through Rate Prediction

Click-through rate (CTR) prediction is one of the most central tasks in online advertising systems. Recent deep learning-based models that exploit feature embedding and high-order data nonlinearity have shown dramatic successes in CTR prediction. However, these models work poorly on cold-start ads with new IDs, whose embeddings are not well learned yet. In this paper, we propose Graph Meta Embedding (GME) models that can rapidly learn how to generate desirable initial embeddings for new ad IDs based on graph neural networks and meta learning. Previous works address this problem from the new ad itself, but ignore possibly useful information contained in existing old ads. In contrast, GMEs simultaneously consider two information sources: the new ad and existing old ads. For the new ad, GMEs exploit its associated attributes. For existing old ads, GMEs first build a graph to connect them with new ads, and then adaptively distill useful information. We propose three specific GMEs from different perspectives to explore what kind of information to use and how to distill information. In particular, GME-P uses Pre-trained neighbor ID embeddings, GME-G uses Generated neighbor ID embeddings and GME-A uses neighbor Attributes. Experimental results on three real-world datasets show that GMEs can significantly improve the prediction performance in both cold-start (i.e., no training data is available) and warm-up (i.e., a small number of training samples are collected) scenarios over five major deep learning-based CTR prediction models. GMEs can be applied to conversion rate (CVR) prediction as well.

[1]  Hugo Larochelle,et al.  A Meta-Learning Perspective on Cold-Start Recommendations for Items , 2017, NIPS.

[2]  Maksims Volkovs,et al.  DropoutNet: Addressing Cold Start in Recommender Systems , 2017, NIPS.

[3]  Jeffrey Dean,et al.  Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.

[4]  Yehuda Koren,et al.  Matrix Factorization Techniques for Recommender Systems , 2009, Computer.

[5]  Cameron Musco,et al.  InfiniteWalk: Deep Network Embeddings as Laplacian Embeddings with a Nonlinearity , 2020, KDD.

[6]  Chang Zhou,et al.  Deep Interest Evolution Network for Click-Through Rate Prediction , 2018, AAAI.

[7]  Shuang-Hong Yang,et al.  Functional matrix factorizations for cold-start recommendation , 2011, SIGIR.

[8]  Amin Mantrach,et al.  Item cold-start recommendations: learning local collective embeddings , 2014, RecSys '14.

[9]  Martin Wattenberg,et al.  Ad click prediction: a view from the trenches , 2013, KDD.

[10]  Li Li,et al.  Deep Spatio-Temporal Neural Networks for Click-Through Rate Prediction , 2019, KDD.

[11]  Jun Wang,et al.  Product-Based Neural Networks for User Response Prediction , 2016, 2016 IEEE 16th International Conference on Data Mining (ICDM).

[12]  Yuan Yu,et al.  TensorFlow: A system for large-scale machine learning , 2016, OSDI.

[13]  Gang Fu,et al.  Deep & Cross Network for Ad Click Predictions , 2017, ADKDD@KDD.

[14]  Guorui Zhou,et al.  Deep Interest Network for Click-Through Rate Prediction , 2017, KDD.

[15]  Nathaniel Good,et al.  Naïve filterbots for robust cold-start recommendations , 2006, KDD '06.

[16]  Xing Xie,et al.  xDeepFM: Combining Explicit and Implicit Feature Interactions for Recommender Systems , 2018, KDD.

[17]  Li Li,et al.  Click-through rate prediction with the user memory network , 2019, Proceedings of the 1st International Workshop on Deep Learning Practice for High-Dimensional Sparse Data.

[18]  Junseok Kwon,et al.  Deep Meta Learning for Real-Time Target-Aware Visual Tracking , 2017, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).

[19]  Chih-Jen Lin,et al.  Field-aware Factorization Machines for CTR Prediction , 2016, RecSys.

[20]  Jie Tang,et al.  Addressing cold start in recommender systems: a semi-supervised co-training algorithm , 2014, SIGIR.

[21]  Xuanjing Huang,et al.  Meta Multi-Task Learning for Sequence Modeling , 2018, AAAI.

[22]  Andrew L. Maas Rectifier Nonlinearities Improve Neural Network Acoustic Models , 2013 .

[23]  Ming Yang,et al.  A Practical Exploration System for Search Advertising , 2017, KDD.

[24]  Steffen Rendle,et al.  Factorization Machines , 2010, 2010 IEEE International Conference on Data Mining.

[25]  Heng-Tze Cheng,et al.  Wide & Deep Learning for Recommender Systems , 2016, DLRS@RecSys.

[26]  Tat-Seng Chua,et al.  Neural Collaborative Filtering , 2017, WWW.

[27]  Yu Sun,et al.  Field-weighted Factorization Machines for Click-Through Rate Prediction in Display Advertising , 2018, WWW.

[28]  Brian D. Davison,et al.  Exploiting contextual factors for click modeling in sponsored search , 2014, WSDM.

[29]  Philippe Preux,et al.  Cold-start Problems in Recommendation Systems via Contextual-bandit Algorithms , 2014, ArXiv.

[30]  Rómer Rosales,et al.  Simple and Scalable Response Prediction for Display Advertising , 2014, ACM Trans. Intell. Syst. Technol..

[31]  Pietro Liò,et al.  Graph Attention Networks , 2017, ICLR.

[32]  Sharath Chandra Guntuku,et al.  Latent Factor Representations for Cold-Start Video Recommendation , 2016, RecSys.

[33]  Yunming Ye,et al.  DeepFM: A Factorization-Machine based Neural Network for CTR Prediction , 2017, IJCAI.

[34]  Hyunsouk Cho,et al.  MeLU: Meta-Learned User Preference Estimator for Cold-Start Recommendation , 2019, KDD.

[35]  Joaquin Quiñonero Candela,et al.  Practical Lessons from Predicting Clicks on Ads at Facebook , 2014, ADKDD'14.

[36]  Naonori Ueda,et al.  Higher-Order Factorization Machines , 2016, NIPS.

[37]  Matthew Richardson,et al.  Predicting clicks: estimating the click-through rate for new ads , 2007, WWW '07.

[38]  Qing He,et al.  Warm Up Cold-start Advertisements: Improving CTR Predictions via Learning to Learn ID Embeddings , 2019, SIGIR.

[39]  Xiuwu Zhang,et al.  Representation Learning-Assisted Click-Through Rate Prediction , 2019, IJCAI.

[40]  Ingrid Zukerman,et al.  Personalised rating prediction for new users using latent factor models , 2011, HT '11.

[41]  Wei Chu,et al.  A contextual-bandit approach to personalized news article recommendation , 2010, WWW '10.

[42]  Liang Tang,et al.  Personalized Recommendation via Parameter-Free Contextual Bandits , 2015, SIGIR.

[43]  Chuan Shi,et al.  Meta-learning on Heterogeneous Information Networks for Cold-start Recommendation , 2020, KDD.

[44]  Sergey Levine,et al.  Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.

[45]  Xiuwu Zhang,et al.  MiNet: Mixed Interest Network for Cross-Domain Click-Through Rate Prediction , 2020, CIKM.

[46]  Weinan Zhang,et al.  User Behavior Retrieval for Click-Through Rate Prediction , 2020, SIGIR.

[47]  Xiao Ma,et al.  Entire Space Multi-Task Model: An Effective Approach for Estimating Post-Click Conversion Rate , 2018, SIGIR.

[48]  Tao Xiang,et al.  Incremental Few-Shot Object Detection , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[49]  Kyunghyun Cho,et al.  Dynamic Meta-Embeddings for Improved Sentence Representations , 2018, EMNLP.

[50]  Yoshua Bengio,et al.  Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.

[51]  Jun Wang,et al.  Deep Learning over Multi-field Categorical Data - - A Case Study on User Response Prediction , 2016, ECIR.

[52]  Jian Tang,et al.  AutoInt: Automatic Feature Interaction Learning via Self-Attentive Neural Networks , 2018, CIKM.

[53]  Philip S. Yu,et al.  Lifelong Domain Word Embedding via Meta-Learning , 2018, IJCAI.

[54]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[55]  Yiming Yang,et al.  Personalized active learning for collaborative filtering , 2008, SIGIR '08.

[56]  Tie-Yan Liu,et al.  Relational click prediction for sponsored search , 2012, WSDM '12.

[57]  Yehuda Koren,et al.  Adaptive bootstrapping of recommender systems using decision trees , 2011, WSDM '11.