A Neural Autoregressive Framework for Collaborative Filtering

Restricted Boltzmann Machine (RBM) is a two layer undirected graph model that capable to represent complex distributions. Recent research has shown RBM-based approach has comparable performance with or even better performance than previous models on many collaborative filtering (CF) tasks. However, the intractable inference makes the training of RBM sophisticated, which prevents it from practical application. We present a novel feedforward neural framework for collaborative filtering called NACF, which is extended from the Neural Autoregressive Distribution Estimator (NADE). Because of the autoregressive feed-forward architecture, NACF can be trained with efficient stochastic gradient descent, instead of slow and hard-to-tune truncated Gibbs sampling for RBM. By introducing linear visible units and dual reversed ordering, NACF show faster convergence and better results than Probabilistic Matrix Factorization (PMF) and corresponding RBM models on MovieLens dataset. Besides, by combining NACF results, the rating prediction of efficientsignificantly improved, showing NACF is an effective and efficient model for collaborative filtering.

[1]  Michael J. Pazzani,et al.  Learning Collaborative Information Filters , 1998, ICML.

[2]  Yan Liu,et al.  Collaborative Topic Regression with Social Matrix Factorization for Recommendation Systems , 2012, ICML.

[3]  Martin Ester,et al.  A matrix factorization technique with trust propagation for recommendation in social networks , 2010, RecSys '10.

[4]  Bong-Jin Yum,et al.  Collaborative filtering based on iterative principal component analysis , 2005, Expert Syst. Appl..

[5]  Alexander J. Smola,et al.  Parallelized Stochastic Gradient Descent , 2010, NIPS.

[6]  Hugo Larochelle,et al.  The Neural Autoregressive Distribution Estimator , 2011, AISTATS.

[7]  Taghi M. Khoshgoftaar,et al.  A Survey of Collaborative Filtering Techniques , 2009, Adv. Artif. Intell..

[8]  Kenneth Y. Goldberg,et al.  Eigentaste: A Constant Time Collaborative Filtering Algorithm , 2001, Information Retrieval.

[9]  James Martens,et al.  Deep learning via Hessian-free optimization , 2010, ICML.

[10]  Preslav Nakov,et al.  A non-IID Framework for Collaborative Filtering with Restricted Boltzmann Machines , 2013, ICML.

[11]  Ruslan Salakhutdinov,et al.  Probabilistic Matrix Factorization , 2007, NIPS.

[12]  Geoffrey E. Hinton A Practical Guide to Training Restricted Boltzmann Machines , 2012, Neural Networks: Tricks of the Trade.

[13]  Hugo Larochelle,et al.  A Deep and Tractable Density Estimator , 2013, ICML.

[14]  Pietro Laface,et al.  Parallel implementation of artificial neural network training , 2010, 2010 IEEE International Conference on Acoustics, Speech and Signal Processing.

[15]  Marc'Aurelio Ranzato,et al.  Large Scale Distributed Deep Networks , 2012, NIPS.

[16]  Geoffrey E. Hinton,et al.  Restricted Boltzmann machines for collaborative filtering , 2007, ICML '07.

[17]  Thomas D. Nielsen,et al.  A latent model for collaborative filtering , 2012, Int. J. Approx. Reason..

[18]  Hugo Larochelle,et al.  A Neural Autoregressive Topic Model , 2012, NIPS.

[19]  Yoshua. Bengio,et al.  Learning Deep Architectures for AI , 2007, Found. Trends Mach. Learn..

[20]  Patrick Seemann,et al.  Matrix Factorization Techniques for Recommender Systems , 2014 .

[21]  John Riedl,et al.  Item-based collaborative filtering recommendation algorithms , 2001, WWW '01.

[22]  Geoffrey E. Hinton,et al.  Reducing the Dimensionality of Data with Neural Networks , 2006, Science.

[23]  Jorge Nocedal,et al.  On the limited memory BFGS method for large scale optimization , 1989, Math. Program..