The Multi-fidelity Multi-armed Bandit
暂无分享,去创建一个
Kirthevasan Kandasamy | Barnabás Póczos | Jeff G. Schneider | Gautam Dasarathy | J. Schneider | Kirthevasan Kandasamy | B. Póczos | Gautam Dasarathy
[1] T. L. Lai Andherbertrobbins. Asymptotically Efficient Adaptive Allocation Rules , 2022 .
[2] Csaba Szepesvári,et al. Exploration-exploitation tradeoff using variance estimates in multi-armed bandits , 2009, Theor. Comput. Sci..
[3] Nenghai Yu,et al. Thompson Sampling for Budgeted Multi-Armed Bandits , 2015, IJCAI.
[4] Ilan Kroo,et al. A Multifidelity Gradient-Free Optimization Method and Application to Aerodynamic Design , 2008 .
[5] Peter Auer,et al. Using Confidence Bounds for Exploitation-Exploration Trade-offs , 2003, J. Mach. Learn. Res..
[6] Larry Wasserman,et al. All of Statistics: A Concise Course in Statistical Inference , 2004 .
[7] R. A. Miller,et al. Sequential kriging optimization using multiple-fidelity evaluations , 2006 .
[8] H. Robbins. Some aspects of the sequential design of experiments , 1952 .
[9] Sébastien Bubeck,et al. Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems , 2012, Found. Trends Mach. Learn..
[10] Jonathan P. How,et al. Reinforcement learning with multi-fidelity simulators , 2014, 2014 IEEE International Conference on Robotics and Automation (ICRA).
[11] W. R. Thompson. ON THE LIKELIHOOD THAT ONE UNKNOWN PROBABILITY EXCEEDS ANOTHER IN VIEW OF THE EVIDENCE OF TWO SAMPLES , 1933 .
[12] R. Agrawal. Sample mean based index policies by O(log n) regret for the multi-armed bandit problem , 1995, Advances in Applied Probability.
[13] Kirthevasan Kandasamy,et al. Gaussian Process Bandit Optimisation with Multi-fidelity Evaluations , 2016, NIPS.
[14] Nicholas R. Jennings,et al. Efficient Regret Bounds for Online Bid Optimisation in Budget-Limited Sponsored Search Auctions , 2014, UAI.
[15] Ran El-Yaniv,et al. Online Choice of Active Learning Algorithms , 2003, J. Mach. Learn. Res..
[16] Kamalika Chaudhuri,et al. Active Learning from Weak and Strong Labelers , 2015, NIPS.