Structured Best Arm Identification with Fixed Confidence
暂无分享,去创建一个
Csaba Szepesvári | Martin Müller | Ruitong Huang | Mohammad M. Ajallooeian | Csaba Szepesvari | Ruitong Huang | Martin Müller
[1] David A. McAllester. Conspiracy Numbers for Min-Max Search , 1988, Artif. Intell..
[2] Alexandre B. Tsybakov,et al. Introduction to Nonparametric Estimation , 2008, Springer series in statistics.
[3] Max Simchowitz,et al. The Simulator: Understanding Adaptive Sampling in the Moderate-Confidence Regime , 2017, COLT.
[4] Aurélien Garivier,et al. Optimal Best Arm Identification with Fixed Confidence , 2016, COLT.
[5] Peter Auer,et al. The Nonstochastic Multiarmed Bandit Problem , 2002, SIAM J. Comput..
[6] Robert D. Nowak,et al. Best-arm identification algorithms for multi-armed bandits in the fixed confidence setting , 2014, 2014 48th Annual Conference on Information Sciences and Systems (CISS).
[7] Ambuj Tewari,et al. PAC Subset Selection in Stochastic Multi-armed Bandits , 2012, ICML.
[8] R. Munos,et al. Best Arm Identification in Multi-Armed Bandits , 2010, COLT.
[9] Jian Li,et al. On the Optimal Sample Complexity for Best Arm Identification , 2015, ArXiv.
[10] Alessandro Lazaric,et al. Best Arm Identification: A Unified Approach to Fixed Budget and Fixed Confidence , 2012, NIPS.
[11] Tor Lattimore,et al. On Explore-Then-Commit strategies , 2016, NIPS.
[12] Michèle Sebag,et al. The grand challenge of computer Go , 2012, Commun. ACM.
[13] L. V. Allis,et al. Searching for solutions in games and artificial intelligence , 1994 .
[14] Demis Hassabis,et al. Mastering the game of Go with deep neural networks and tree search , 2016, Nature.
[15] Wouter M. Koolen,et al. Monte-Carlo Tree Search by Best Arm Identification , 2017, NIPS.
[16] Shie Mannor,et al. Action Elimination and Stopping Conditions for the Multi-Armed Bandit and Reinforcement Learning Problems , 2006, J. Mach. Learn. Res..
[17] Eiji Takimoto,et al. Efficient Sampling Method for Monte Carlo Tree Search Problem , 2014, IEICE Trans. Inf. Syst..
[18] Matthew Malloy,et al. lil' UCB : An Optimal Exploration Algorithm for Multi-Armed Bandits , 2013, COLT.
[19] Wouter M. Koolen,et al. Maximin Action Identification: A New Bandit Framework for Games , 2016, COLT.
[20] Oren Somekh,et al. Almost Optimal Exploration in Multi-Armed Bandits , 2013, ICML.
[21] Aurélien Garivier,et al. On the Complexity of Best-Arm Identification in Multi-Armed Bandit Models , 2014, J. Mach. Learn. Res..