暂无分享,去创建一个
Nello Cristianini | Rémi Munos | Michal Valko | Nathaniel Korda | Ilias N. Flaounas | R. Munos | N. Cristianini | N. Korda | Michal Valko | I. Flaounas
[1] Peter Auer,et al. Using Confidence Bounds for Exploitation-Exploration Trade-offs , 2003, J. Mach. Learn. Res..
[2] Wei Chu,et al. A contextual-bandit approach to personalized news article recommendation , 2010, WWW '10.
[3] J. Langford,et al. The Epoch-Greedy algorithm for contextual multi-armed bandits , 2007, NIPS 2007.
[4] T. L. Lai Andherbertrobbins. Asymptotically Efficient Adaptive Allocation Rules , 2022 .
[5] John Langford,et al. Contextual Bandit Algorithms with Supervised Learning Guarantees , 2010, AISTATS.
[6] Martin Pál,et al. Contextual Multi-Armed Bandits , 2010, AISTATS.
[7] Nello Cristianini,et al. Kernel Methods for Pattern Analysis , 2003, ICTAI.
[8] John Langford,et al. The Epoch-Greedy Algorithm for Multi-armed Bandits with Side Information , 2007, NIPS.
[9] Sébastien Bubeck,et al. Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems , 2012, Found. Trends Mach. Learn..
[10] John Langford,et al. Efficient Optimal Learning for Contextual Bandits , 2011, UAI.
[11] Bruno Pouliquen,et al. An introduction to the Europe Media Monitor family of applications , 2013, ArXiv.
[12] John N. Tsitsiklis,et al. Linearly Parameterized Bandits , 2008, Math. Oper. Res..
[13] Wei Chu,et al. Contextual Bandits with Linear Payoff Functions , 2011, AISTATS.
[14] Maya R. Gupta,et al. Similarity-based Classification: Concepts and Algorithms , 2009, J. Mach. Learn. Res..
[15] John Shawe-Taylor,et al. Regret Bounds for Gaussian Process Bandit Problems , 2010, AISTATS 2010.
[16] Andreas Krause,et al. Contextual Gaussian Process Bandit Optimization , 2011, NIPS.
[17] Eli Upfal,et al. Multi-Armed Bandits in Metric Spaces ∗ , 2008 .
[18] Csaba Szepesvári,et al. Improved Algorithms for Linear Stochastic Bandits , 2011, NIPS.
[19] Aleksandrs Slivkins,et al. Contextual Bandits with Similarity Information , 2009, COLT.
[20] John Shawe-Taylor,et al. PAC-Bayesian Analysis of Contextual Bandits , 2011, NIPS.
[21] Peter Auer,et al. The Nonstochastic Multiarmed Bandit Problem , 2002, SIAM J. Comput..
[22] Gábor Lugosi,et al. Prediction, learning, and games , 2006 .
[23] Fuzhen Zhang. The Schur complement and its applications , 2005 .
[24] Thomas P. Hayes,et al. Stochastic Linear Optimization under Bandit Feedback , 2008, COLT.
[25] Andreas Krause,et al. Information-Theoretic Regret Bounds for Gaussian Process Optimization in the Bandit Setting , 2009, IEEE Transactions on Information Theory.