Linear Stochastic Bandits Under Safety Constraints
暂无分享,去创建一个
Christos Thrampoulidis | Mahnoosh Alizadeh | Sanae Amani | M. Alizadeh | Christos Thrampoulidis | Sanae Amani
[1] Jaime F. Fisac,et al. Reachability-based safe learning with Gaussian processes , 2014, 53rd IEEE Conference on Decision and Control.
[2] Duy Nguyen-Tuong,et al. Safe Exploration for Active Learning with Gaussian Processes , 2015, ECML/PKDD.
[3] Benjamin Van Roy,et al. Learning to Optimize via Posterior Sampling , 2013, Math. Oper. Res..
[4] Pieter Abbeel,et al. Safe Exploration in Markov Decision Processes , 2012, ICML.
[5] Alkis Gotovos,et al. Safe Exploration for Optimization with Gaussian Processes , 2015, ICML.
[6] Joel A. Tropp,et al. An Introduction to Matrix Concentration Inequalities , 2015, Found. Trends Mach. Learn..
[7] Yifan Wu,et al. Conservative Bandits , 2016, ICML.
[8] S. Shankar Sastry,et al. Provably safe and robust learning-based model predictive control , 2011, Autom..
[9] Aurélien Garivier,et al. Parametric Bandits: The Generalized Linear Case , 2010, NIPS.
[10] Thomas P. Hayes,et al. Stochastic Linear Optimization under Bandit Feedback , 2008, COLT.
[11] Andreas Krause,et al. Information-Theoretic Regret Bounds for Gaussian Process Optimization in the Bandit Setting , 2009, IEEE Transactions on Information Theory.
[12] Csaba Szepesvári,et al. Improved Algorithms for Linear Stochastic Bandits , 2011, NIPS.
[13] Claire J. Tomlin,et al. Guaranteed safe online learning of a bounded system , 2011, 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems.
[14] John Langford,et al. Resourceful Contextual Bandits , 2014, COLT.
[15] Pieter Abbeel,et al. Constrained Policy Optimization , 2017, ICML.
[16] Peter Auer,et al. Finite-time Analysis of the Multiarmed Bandit Problem , 2002, Machine Learning.
[17] Andreas Krause,et al. Bayesian optimization with safety constraints: safe and automatic parameter tuning in robotics , 2016, Machine Learning.
[18] Nikhil R. Devanur,et al. Linear Contextual Bandits with Knapsacks , 2015, NIPS.
[19] Joel W. Burdick,et al. Stagewise Safe Bayesian Optimization with Gaussian Processes , 2018, ICML.
[20] Andreas Krause,et al. Safe Convex Learning under Uncertain Constraints , 2019, AISTATS.
[21] Lihong Li,et al. Provable Optimal Algorithms for Generalized Linear Contextual Bandits , 2017, ArXiv.
[22] Aleksandrs Slivkins,et al. Bandits with Knapsacks , 2013, 2013 IEEE 54th Annual Symposium on Foundations of Computer Science.
[23] Csaba Szepesvári,et al. Exploration-exploitation tradeoff using variance estimates in multi-armed bandits , 2009, Theor. Comput. Sci..
[24] Angela P. Schoellig,et al. Robust Constrained Learning-based NMPC enabling reliable mobile robot path tracking , 2016, Int. J. Robotics Res..
[25] John N. Tsitsiklis,et al. Linearly Parameterized Bandits , 2008, Math. Oper. Res..
[26] Benjamin Van Roy,et al. Conservative Contextual Linear Bandits , 2016, NIPS.
[27] Wei Chu,et al. Contextual Bandits with Linear Payoff Functions , 2011, AISTATS.
[28] R. Srikant,et al. Algorithms with Logarithmic or Sublinear Regret for Constrained Contextual Bandits , 2015, NIPS.