暂无分享,去创建一个
[1] Karthik Sridharan,et al. ZigZag: A New Approach to Adaptive Online Learning , 2017, COLT.
[2] Yoav Freund,et al. A decision-theoretic generalization of on-line learning and an application to boosting , 1997, EuroCOLT.
[3] Francesco Orabona. A Modern Introduction to Online Learning , 2019, ArXiv.
[4] Alessandro Lazaric,et al. Exploiting easy data in online optimization , 2014, NIPS.
[5] Haipeng Luo,et al. More Adaptive Algorithms for Adversarial Bandits , 2018, COLT.
[6] Julian Zimmert,et al. Tsallis-INF: An Optimal Algorithm for Stochastic and Adversarial Bandits , 2018, J. Mach. Learn. Res..
[7] Sanjeev Arora,et al. The Multiplicative Weights Update Method: a Meta-Algorithm and Applications , 2012, Theory Comput..
[8] Karthik Sridharan,et al. Adaptive Online Learning , 2015, NIPS.
[9] Aleksandrs Slivkins,et al. Corruption Robust Exploration in Episodic Reinforcement Learning , 2019, ArXiv.
[10] Elad Hazan,et al. Introduction to Online Convex Optimization , 2016, Found. Trends Optim..
[11] Yishay Mansour,et al. Improved second-order bounds for prediction with expert advice , 2006, Machine Learning.
[12] Stéphane Gaïffas,et al. On the optimality of the Hedge algorithm in the stochastic regime , 2018, J. Mach. Learn. Res..
[13] Lihong Li,et al. Adversarial Attacks on Stochastic Bandits , 2018, NeurIPS.
[14] Wouter M. Koolen,et al. Learning the Learning Rate for Prediction with Expert Advice , 2014, NIPS.
[15] Purushottam Kar,et al. Corruption-tolerant bandit learning , 2018, Machine Learning.
[16] Rong Jin,et al. 25th Annual Conference on Learning Theory Online Optimization with Gradual Variations , 2022 .
[17] Yoav Freund,et al. A decision-theoretic generalization of on-line learning and an application to boosting , 1995, EuroCOLT.
[18] Claudio Gentile,et al. Adaptive and Self-Confident On-Line Learning Algorithms , 2000, J. Comput. Syst. Sci..
[19] Ness Shroff,et al. Data Poisoning Attacks on Stochastic Bandits , 2019, ICML.
[20] Karthik Sridharan,et al. Online Learning with Predictable Sequences , 2012, COLT.
[21] Manfred K. Warmuth,et al. The Weighted Majority Algorithm , 1994, Inf. Comput..
[22] Wouter M. Koolen,et al. Adaptive Hedge , 2011, NIPS.
[23] Anupam Gupta,et al. Better Algorithms for Stochastic Bandits with Adversarial Corruptions , 2019, COLT.
[24] Elad Hazan,et al. Extracting certainty from uncertainty: regret bounded by variation in costs , 2008, Machine Learning.
[25] Renato Paes Leme,et al. Stochastic bandits robust to adversarial corruptions , 2018, STOC.
[26] Wouter M. Koolen,et al. MetaGrad: Multiple Learning Rates in Online Learning , 2016, NIPS.
[27] Wouter M. Koolen,et al. Follow the leader if you can, hedge if you must , 2013, J. Mach. Learn. Res..
[28] Wouter M. Koolen,et al. Second-order Quantile Methods for Experts and Combinatorial Games , 2015, COLT.
[29] Yoram Singer,et al. Adaptive Subgradient Methods for Online Learning and Stochastic Optimization , 2011, J. Mach. Learn. Res..
[30] Mark D. Reid,et al. Fast rates in statistical and online learning , 2015, J. Mach. Learn. Res..
[31] Francesco Orabona,et al. Scale-free online learning , 2016, Theor. Comput. Sci..
[32] Wouter M. Koolen,et al. Combining Adversarial Guarantees and Stochastic Fast Rates in Online Learning , 2016, NIPS.
[33] Shai Shalev-Shwartz,et al. Online Learning and Online Convex Optimization , 2012, Found. Trends Mach. Learn..
[34] Gábor Lugosi,et al. Prediction, learning, and games , 2006 .