PRTNets: Cold-Start Recommendations Using Pairwise Ranking and Transfer Networks

In collaborative filtering, matrix factorization, which decomposes the ratings matrix into low rank user and item latent matrices is widely used. The decomposition is based on the rating scores of users to item, with the user and item latent matrices sharing a common embedding space. A similarity function between the two represents the predicted rating of a user to an item. However, this matrix factorization approach falls short for cold-start recommendation where items have very few or no ratings. This paper puts forward a novel approach of doing cold-start recommendation by using a neural network, the Transfer Network, to learn a nonlinear mapping from item features to the item latent matrix. The item latent matrix is produced by another network, the Pairwise Ranking Network, which utilizes pairwise ranking functions. The Pairwise Ranking Network efficiently utilizes implicit feedback by optimizing the ranking of the recommendation list. We find the optimal architecture for the Pairwise Network and the Transfer Network through warm-start and cold-start evaluation. With the Transfer Network, we map the Tag Genome dataset to the item latent matrix and produce cold-start recommendations for a test set derived from the MovieLens 20M dataset. Our approach yielded a significant margin of improvement of 0.276 and 0.089 average precision at \(k=10\) over the baseline LightFM and neighborhood averaging methods respectively.

[1]  Domonkos Tikk,et al.  Alternating least squares for personalized ranking , 2012, RecSys.

[2]  F. Maxwell Harper,et al.  The MovieLens Datasets: History and Context , 2016, TIIS.

[3]  Benjamin Schrauwen,et al.  Deep content-based music recommendation , 2013, NIPS.

[4]  Yifan Hu,et al.  Collaborative Filtering for Implicit Feedback Datasets , 2008, 2008 Eighth IEEE International Conference on Data Mining.

[5]  Nitish Srivastava,et al.  Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..

[6]  Yehuda Koren,et al.  The BellKor Solution to the Netflix Grand Prize , 2009 .

[7]  Jason Weston,et al.  WSABIE: Scaling Up to Large Vocabulary Image Annotation , 2011, IJCAI.

[8]  Lars Schmidt-Thieme,et al.  Learning Attribute-to-Feature Mappings for Cold-Start Recommendations , 2010, 2010 IEEE International Conference on Data Mining.

[9]  Sergey Ioffe,et al.  Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.

[10]  Premkumar Natarajan,et al.  WMRB: Learning to Rank in a Scalable Batch Training Approach , 2017, RecSys Posters.

[11]  Steffen Rendle,et al.  Factorization Machines , 2010, 2010 IEEE International Conference on Data Mining.

[12]  Maciej Kula,et al.  Metadata Embeddings for User and Item Cold-start Recommendations , 2015, CBRecSys@RecSys.

[13]  Lars Schmidt-Thieme,et al.  BPR: Bayesian Personalized Ranking from Implicit Feedback , 2009, UAI.

[14]  John Riedl,et al.  The Tag Genome: Encoding Community Knowledge to Support Novel Interaction , 2012, TIIS.