暂无分享,去创建一个
Olivier Sigaud | Alexandre Laterre | Karim Beguir | Thomas Pierrot | Louis Monier | Jakub Kmec | Valentin Courgeau | Olivier Sigaud | L. Monier | Valentin Courgeau | Alexandre Laterre | Karim Beguir | Thomas Pierrot | Jakub Kmec
[1] Justin Fu,et al. D4RL: Datasets for Deep Data-Driven Reinforcement Learning , 2020, ArXiv.
[2] Olivier Sigaud,et al. QD-RL: Efficient Mixing of Quality and Diversity in Reinforcement Learning , 2020, ArXiv.
[3] Nando de Freitas,et al. RL Unplugged: Benchmarks for Offline Reinforcement Learning , 2020, ArXiv.
[4] George Tucker,et al. Conservative Q-Learning for Offline Reinforcement Learning , 2020, NeurIPS.
[5] Herke van Hoof,et al. Addressing Function Approximation Error in Actor-Critic Methods , 2018, ICML.
[6] Sergey Levine,et al. Stabilizing Off-Policy Q-Learning via Bootstrapping Error Reduction , 2019, NeurIPS.
[7] Martin A. Riedmiller,et al. Keep Doing What Worked: Behavioral Modelling Priors for Offline Reinforcement Learning , 2020, ICLR.
[8] Seyed Kamyar Seyed Ghasemipour,et al. EMaQ: Expected-Max Q-Learning Operator for Simple Yet Effective Offline and Online RL , 2020, ICML.
[9] Joelle Pineau,et al. Benchmarking Batch Deep Reinforcement Learning Algorithms , 2019, ArXiv.
[10] Sergey Levine,et al. Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor , 2018, ICML.
[11] Nando de Freitas,et al. Critic Regularized Regression , 2020, NeurIPS.
[12] Alex Graves,et al. Playing Atari with Deep Reinforcement Learning , 2013, ArXiv.
[13] Sergey Levine,et al. Accelerating Online Reinforcement Learning with Offline Datasets , 2020, ArXiv.
[14] Sergey Levine,et al. Offline Reinforcement Learning: Tutorial, Review, and Perspectives on Open Problems , 2020, ArXiv.
[15] Doina Precup,et al. Off-Policy Deep Reinforcement Learning without Exploration , 2018, ICML.