Pure Exploration in Multi-armed Bandits Problems
暂无分享,去创建一个
[1] W. Hoeffding. Probability Inequalities for sums of Bounded Random Variables , 1963 .
[2] Colin McDiarmid,et al. Surveys in Combinatorics, 1989: On the method of bounded differences , 1989 .
[3] Luc Devroye,et al. Combinatorial methods in density estimation , 2001, Springer series in statistics.
[4] Y. Freund,et al. The non-stochastic multi-armed bandit problem , 2001 .
[5] Peter Auer,et al. The Nonstochastic Multiarmed Bandit Problem , 2002, SIAM J. Comput..
[6] Shie Mannor,et al. PAC Bounds for Multi-armed Bandit and Markov Decision Processes , 2002, COLT.
[7] John N. Tsitsiklis,et al. The Sample Complexity of Exploration in the Multi-Armed Bandit Problem , 2004, J. Mach. Learn. Res..
[8] Peter Auer,et al. Finite-time Analysis of the Multiarmed Bandit Problem , 2002, Machine Learning.
[9] Russell Greiner,et al. The Budgeted Multi-armed Bandit Problem , 2004, COLT.
[10] Robert D. Kleinberg. Nearly Tight Bounds for the Continuum-Armed Bandit Problem , 2004, NIPS.
[11] Olivier Teytaud,et al. Modification of UCT with Patterns in Monte-Carlo Go , 2006 .
[12] K. Schlag. ELEVEN - Tests needed for a Recommendation , 2006 .
[13] Csaba Szepesvári,et al. Bandit Based Monte-Carlo Planning , 2006, ECML.
[14] Neil D. Lawrence,et al. Missing Data in Kernel PCA , 2006, ECML.
[15] Rémi Munos,et al. Bandit Algorithms for Tree Search , 2007, UAI.
[16] H. Robbins. Some aspects of the sequential design of experiments , 1952 .
[17] Csaba Szepesvári,et al. Online Optimization in X-Armed Bandits , 2008, NIPS.
[18] Rémi Munos,et al. Pure Exploration for Multi-Armed Bandit Problems , 2008, ArXiv.
[19] R. Bass,et al. Review: P. Billingsley, Convergence of probability measures , 1971 .
[20] A. Soifer. Open Problems Session , 2011 .
[21] T. L. Lai Andherbertrobbins. Asymptotically Efficient Adaptive Allocation Rules , 2022 .