A Contextual Bandit Approach to Personalized Online Recommendation via Sparse Interactions
暂无分享,去创建一个
Yang Gao | Hao Wang | Shangdong Yang | Chenyu Zhang | Yang Gao | Chenyu Zhang | Hao Wang | Shangdong Yang
[1] Wei Chu,et al. Contextual Bandits with Linear Payoff Functions , 2011, AISTATS.
[2] Shipra Agrawal,et al. Thompson Sampling for Contextual Bandits with Linear Payoffs , 2012, ICML.
[3] Nick Craswell,et al. An experimental comparison of click position-bias models , 2008, WSDM '08.
[4] Wei Chu,et al. A contextual-bandit approach to personalized news article recommendation , 2010, WWW '10.
[5] J. Tukey,et al. An algorithm for the machine calculation of complex Fourier series , 1965 .
[6] Jiahui Liu,et al. Personalized news recommendation based on click behavior , 2010, IUI '10.
[7] Shuai Li,et al. Collaborative Filtering Bandits , 2015, SIGIR.
[8] Weiwei Xia,et al. A Temporal Item-Based Collaborative Filtering Approach , 2011, FGIT-SIP.
[9] Wei Chu,et al. Unbiased offline evaluation of contextual-bandit-based news article recommendation algorithms , 2010, WSDM '11.
[10] Peter Auer,et al. Using Confidence Bounds for Exploitation-Exploration Trade-offs , 2003, J. Mach. Learn. Res..
[11] Zheng Wen,et al. Cascading Bandits: Learning to Rank in the Cascade Model , 2015, ICML.
[12] Csaba Szepesvári,et al. Improved Algorithms for Linear Stochastic Bandits , 2011, NIPS.
[13] Shuai Li,et al. On Context-Dependent Clustering of Bandits , 2016, ICML.
[14] Peter Auer,et al. Finite-time Analysis of the Multiarmed Bandit Problem , 2002, Machine Learning.