Stochastic Rank-1 Bandits
暂无分享,去创建一个
Zheng Wen | Csaba Szepesvári | Branislav Kveton | Claire Vernade | Sumeet Katariya | Csaba Szepesvari | B. Kveton | Zheng Wen | S. Katariya | Claire Vernade
[1] T. L. Graves,et al. Asymptotically Efficient Adaptive Choice of Control Laws inControlled Markov Chains , 1997 .
[2] Peter Auer,et al. Finite-time Analysis of the Multiarmed Bandit Problem , 2002, Machine Learning.
[3] Eric Moulines,et al. Online EM Algorithm for Latent Data Models , 2007, ArXiv.
[4] Matthew Richardson,et al. Predicting clicks: estimating the click-through rate for new ads , 2007, WWW '07.
[5] Nick Craswell,et al. An experimental comparison of click position-bias models , 2008, WSDM '08.
[6] Thomas P. Hayes,et al. Stochastic Linear Optimization under Bandit Feedback , 2008, COLT.
[7] Andrea Montanari,et al. Matrix Completion from Noisy Entries , 2009, J. Mach. Learn. Res..
[8] Yehuda Koren,et al. Matrix Factorization Techniques for Recommender Systems , 2009, Computer.
[9] Yong Yu,et al. Robust Subspace Segmentation by Low-Rank Representation , 2010, ICML.
[10] Aurélien Garivier,et al. Parametric Bandits: The Generalized Linear Case , 2010, NIPS.
[11] Peter Auer,et al. UCB revisited: Improved regret bounds for the stochastic multi-armed bandit problem , 2010, Period. Math. Hung..
[12] Csaba Szepesvári,et al. Improved Algorithms for Linear Stochastic Bandits , 2011, NIPS.
[13] Jun Wang,et al. Interactive collaborative filtering , 2013, CIKM.
[14] Igal Sason,et al. Concentration of Measure Inequalities in Information Theory, Communications, and Coding , 2012, Found. Trends Commun. Inf. Theory.
[15] Shie Mannor,et al. Latent Bandits , 2014, ICML.
[16] Rémi Munos,et al. Spectral Bandits for Smooth Graph Functions , 2014, ICML.
[17] Shuai Li,et al. Online Clustering of Bandits , 2014, ICML.
[18] Patrick Seemann,et al. Matrix Factorization Techniques for Recommender Systems , 2014 .
[19] Rémi Munos,et al. Spectral Thompson Sampling , 2014, AAAI.
[20] Zheng Wen,et al. Cascading Bandits: Learning to Rank in the Cascade Model , 2015, ICML.
[21] Wtt Wtt. Tight Regret Bounds for Stochastic Combinatorial Semi-Bandits , 2015 .
[22] Long Tran-Thanh,et al. Efficient Thompson Sampling for Online Matrix-Factorization Recommendation , 2015, NIPS.
[23] Zheng Wen,et al. Combinatorial Cascading Bandits , 2015, NIPS.
[24] Alexandre Proutière,et al. Learning to Rank , 2015, SIGMETRICS.
[25] M. de Rijke,et al. Click Models for Web Search , 2015, Click Models for Web Search.
[26] Shuai Li,et al. Contextual Combinatorial Cascading Bandits , 2016, ICML.
[27] Aurélien Garivier,et al. On the Complexity of Best-Arm Identification in Multi-Armed Bandit Models , 2014, J. Mach. Learn. Res..
[28] Shuai Li,et al. Collaborative Filtering Bandits , 2015, SIGIR.
[29] Zheng Wen,et al. DCM Bandits: Learning to Rank with Multiple Clicks , 2016, ICML.
[30] Olivier Cappé,et al. Multiple-Play Bandits in the Position-Based Model , 2016, NIPS.
[31] Zheng Wen,et al. Cascading Bandits for Large-Scale Recommendation Problems , 2016, UAI.
[32] T. L. Lai Andherbertrobbins. Asymptotically Efficient Adaptive Allocation Rules , 2022 .