暂无分享,去创建一个
Hongyi Zhou | Baiming Chen | Zuxin Liu | Ding Zhao | Martial Hebert | Sicheng Zhong | M. Hebert | Ding Zhao | Baiming Chen | Zuxin Liu | Sicheng Zhong | Hongyi Zhou
[1] Sergey Levine,et al. Neural Network Dynamics for Model-Based Deep Reinforcement Learning with Model-Free Fine-Tuning , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).
[2] Torsten Koller,et al. Learning-based Model Predictive Control for Safe Exploration and Reinforcement Learning , 2019, ArXiv.
[3] Ofir Nachum,et al. A Lyapunov-based Approach to Safe Reinforcement Learning , 2018, NeurIPS.
[4] Wenhao Ding,et al. Task-Agnostic Online Reinforcement Learning with an Infinite Mixture of Gaussian Processes , 2020, NeurIPS.
[5] Pieter Abbeel,et al. Responsive Safety in Reinforcement Learning by PID Lagrangian Methods , 2020, ICML.
[6] Zoubin Ghahramani,et al. Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning , 2015, ICML.
[7] Baiming Chen,et al. Delay-Aware Multi-Agent Reinforcement Learning , 2020, ArXiv.
[8] Sergey Levine,et al. Trust Region Policy Optimization , 2015, ICML.
[9] Sergey Levine,et al. Can Autonomous Vehicles Identify, Recover From, and Adapt to Distribution Shifts? , 2020, ICML.
[10] Javier García,et al. A comprehensive survey on safe reinforcement learning , 2015, J. Mach. Learn. Res..
[11] Fritz Wysotzki,et al. Risk-Sensitive Reinforcement Learning Applied to Control under Constraints , 2005, J. Artif. Intell. Res..
[12] Chris Gaskett,et al. Reinforcement learning under circumstances beyond its control , 2003 .
[13] Dirk P. Kroese,et al. Chapter 3 – The Cross-Entropy Method for Optimization , 2013 .
[14] Yuval Tassa,et al. Safe Exploration in Continuous Action Spaces , 2018, ArXiv.
[15] Yanan Sui,et al. Safe Reinforcement Learning in Constrained Markov Decision Processes , 2020, ICML.
[16] Andrew K. C. Wong,et al. Classification of Imbalanced Data: a Review , 2009, Int. J. Pattern Recognit. Artif. Intell..
[17] Pieter Abbeel,et al. Constrained Policy Optimization , 2017, ICML.
[18] Dirk P. Kroese,et al. The cross-entropy method for estimation , 2013 .
[19] Mohammad Ghavamzadeh,et al. Lyapunov-based Safe Policy Optimization for Continuous Control , 2019, ArXiv.
[20] Joelle Pineau,et al. Benchmarking Batch Deep Reinforcement Learning Algorithms , 2019, ArXiv.
[21] Dario Amodei,et al. Benchmarking Safe Exploration in Deep Reinforcement Learning , 2019 .
[22] Tie-Yan Liu,et al. LightGBM: A Highly Efficient Gradient Boosting Decision Tree , 2017, NIPS.
[23] Hongyi Zhou,et al. MAPPER: Multi-Agent Path Planning with Evolutionary Reinforcement Learning in Mixed Dynamic Environments , 2020, 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).
[24] Shie Mannor,et al. Scaling Up Robust MDPs by Reinforcement Learning , 2013, ArXiv.
[25] Wei Wu,et al. Dynamic Curriculum Learning for Imbalanced Data Classification , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[26] Giovanni De Magistris,et al. OptLayer - Practical Constrained Optimization for Deep Reinforcement Learning in the Real World , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).
[27] Baiming Chen,et al. Delay-Aware Multi-Agent Reinforcement Learning , 2020, ArXiv.
[28] Sergey Levine,et al. Deep Reinforcement Learning in a Handful of Trials using Probabilistic Dynamics Models , 2018, NeurIPS.
[29] James M. Rehg,et al. Aggressive Deep Driving: Combining Convolutional Neural Networks and Model Predictive Control , 2017, CoRL.
[30] Tadahiro Taniguchi,et al. Variational Inference MPC for Bayesian Model-based Reinforcement Learning , 2019, CoRL.
[31] Alex Graves,et al. Playing Atari with Deep Reinforcement Learning , 2013, ArXiv.
[32] Alkis Gotovos,et al. Safe Exploration for Optimization with Gaussian Processes , 2015, ICML.
[33] Constrained Cross-Entropy Method for Safe Reinforcement Learning , 2018 .
[34] Andreas Krause,et al. Safe Model-based Reinforcement Learning with Stability Guarantees , 2017, NIPS.
[35] Eitan Altman,et al. Constrained Markov decision processes with total cost criteria: Lagrangian approach and dual linear program , 1998, Math. Methods Oper. Res..