暂无分享,去创建一个
[1] P. Auer,et al. Adaptively Tracking the Best Arm with an Unknown Number of Distribution Changes , 2018 .
[2] Haipeng Luo,et al. Efficient Contextual Bandits in Non-stationary Worlds , 2017, COLT.
[3] Eli Upfal,et al. Adapting to a Changing Environment: the Brownian Restless Bandits , 2008, COLT.
[4] Haipeng Luo,et al. A New Algorithm for Non-stationary Contextual Bandits: Efficient, Optimal, and Parameter-free , 2019, COLT.
[5] Aurélien Garivier,et al. Optimal Best Arm Identification with Fixed Confidence , 2016, COLT.
[6] Omar Besbes,et al. Optimal Exploration-Exploitation in a Multi-Armed-Bandit Problem with Non-Stationary Rewards , 2014, Stochastic Systems.
[7] Aurélien Garivier,et al. Explore First, Exploit Next: The True Shape of Regret in Bandit Problems , 2016, Math. Oper. Res..
[8] David Tse,et al. Fundamentals of Wireless Communication , 2005 .
[9] Zohar S. Karnin,et al. Multi-armed Bandits: Competing with Optimal Sequences , 2016, NIPS.
[10] Eric Moulines,et al. On Upper-Confidence Bound Policies for Switching Bandit Problems , 2011, ALT.
[11] Peter Auer,et al. The Nonstochastic Multiarmed Bandit Problem , 2002, SIAM J. Comput..
[12] Gábor Lugosi,et al. Prediction, learning, and games , 2006 .
[13] Yuchen Lu,et al. Countering Language Drift with Seeded Iterated Learning , 2020, ICML.
[14] Peter Auer,et al. Adaptively Tracking the Best Bandit Arm with an Unknown Number of Distribution Changes , 2019, COLT.
[15] Omar Besbes,et al. Stochastic Multi-Armed-Bandit Problem with Non-stationary Rewards , 2014, NIPS.
[16] Aleksandrs Slivkins,et al. Introduction to Multi-Armed Bandits , 2019, Found. Trends Mach. Learn..