暂无分享,去创建一个
[1] Philip Bachman,et al. Deep Reinforcement Learning that Matters , 2017, AAAI.
[2] Glen Berseth,et al. DeepLoco , 2017, ACM Trans. Graph..
[3] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[4] Peter Henderson,et al. Reproducibility of Benchmarked Deep Reinforcement Learning Tasks for Continuous Control , 2017, ArXiv.
[5] Elliot Meyerson,et al. Evolving Deep Neural Networks , 2017, Artificial Intelligence in the Age of Neural Networks and Brain Computing.
[6] Sergey Levine,et al. Learning to Walk in the Real World with Minimal Human Effort , 2020, CoRL.
[7] Jason Yosinski,et al. Deconstructing Lottery Tickets: Zeros, Signs, and the Supermask , 2019, NeurIPS.
[8] Lydia Tapia,et al. PRM-RL: Long-range Robotic Navigation Tasks by Combining Reinforcement Learning and Sampling-Based Planning , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).
[9] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[10] Yuval Tassa,et al. MuJoCo: A physics engine for model-based control , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.
[11] Sergey Levine,et al. Uncertainty-Aware Reinforcement Learning for Collision Avoidance , 2017, ArXiv.
[12] Alex Kendall,et al. Concrete Dropout , 2017, NIPS.
[13] Adam Gaier,et al. Weight Agnostic Neural Networks , 2019, NeurIPS.
[14] Joonho Lee,et al. DeepGait: Planning and Control of Quadrupedal Gaits Using Deep Reinforcement Learning , 2020, IEEE Robotics and Automation Letters.
[15] Zoubin Ghahramani,et al. Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning , 2015, ICML.
[16] Risto Miikkulainen,et al. Evolving Neural Networks through Augmenting Topologies , 2002, Evolutionary Computation.
[17] Rob Fergus,et al. Visualizing and Understanding Convolutional Networks , 2013, ECCV.
[18] Michael Carbin,et al. The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks , 2018, ICLR.
[19] Katia P. Sycara,et al. Transparency and Explanation in Deep Reinforcement Learning Neural Networks , 2018, AIES.
[20] Marcin Andrychowicz,et al. What Matters In On-Policy Reinforcement Learning? A Large-Scale Empirical Study , 2020, ArXiv.
[21] Gregory J. Wolff,et al. Optimal Brain Surgeon and general network pruning , 1993, IEEE International Conference on Neural Networks.
[22] Sergey Levine,et al. Learning to Walk via Deep Reinforcement Learning , 2018, Robotics: Science and Systems.
[23] Daniel D. Lee,et al. Efficient learning of stand-up motion for humanoid robots with bilateral symmetry , 2016, 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).
[24] Alex Graves,et al. Asynchronous Methods for Deep Reinforcement Learning , 2016, ICML.
[25] Aleksandra Faust,et al. Learning Navigation Behaviors End-to-End With AutoRL , 2018, IEEE Robotics and Automation Letters.
[26] Sergey Levine,et al. DeepMimic , 2018, ACM Trans. Graph..
[27] Michiel van de Panne,et al. Learning locomotion skills using DeepRL: does the choice of action space matter? , 2016, Symposium on Computer Animation.
[28] Wojciech Zaremba,et al. OpenAI Gym , 2016, ArXiv.
[29] Bohan Wu,et al. MAT: Multi-Fingered Adaptive Tactile Grasping via Deep Reinforcement Learning , 2019, CoRL.
[30] Sangbae Kim,et al. Mini Cheetah: A Platform for Pushing the Limits of Dynamic Quadruped Control , 2019, 2019 International Conference on Robotics and Automation (ICRA).
[31] Quoc V. Le,et al. Neural Architecture Search with Reinforcement Learning , 2016, ICLR.
[32] Tariq Samad,et al. Designing Application-Specific Neural Networks Using the Genetic Algorithm , 1989, NIPS.
[33] Mingjie Sun,et al. Rethinking the Value of Network Pruning , 2018, ICLR.
[34] Yuval Tassa,et al. Continuous control with deep reinforcement learning , 2015, ICLR.
[35] Alec Radford,et al. Proximal Policy Optimization Algorithms , 2017, ArXiv.
[36] Joonho Lee,et al. Learning agile and dynamic motor skills for legged robots , 2019, Science Robotics.
[37] Philip H. S. Torr,et al. SNIP: Single-shot Network Pruning based on Connection Sensitivity , 2018, ICLR.
[38] Yuval Tassa,et al. DeepMind Control Suite , 2018, ArXiv.
[39] Youngchul Sung,et al. Message-Dropout: An Efficient Training Method for Multi-Agent Deep Reinforcement Learning , 2019, AAAI.
[40] Herke van Hoof,et al. Addressing Function Approximation Error in Actor-Critic Methods , 2018, ICML.
[41] Alok Aggarwal,et al. Regularized Evolution for Image Classifier Architecture Search , 2018, AAAI.
[42] Song Han,et al. Learning both Weights and Connections for Efficient Neural Network , 2015, NIPS.
[43] Michiel van de Panne,et al. Learning to Locomote: Understanding How Environment Design Matters for Deep Reinforcement Learning , 2020, MIG.
[44] Yann LeCun,et al. Optimal Brain Damage , 1989, NIPS.
[45] C. Rasmussen,et al. Improving PILCO with Bayesian Neural Network Dynamics Models , 2016 .
[46] Sergey Levine,et al. Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor , 2018, ICML.
[47] Jonathan Dodge,et al. Visualizing and Understanding Atari Agents , 2017, ICML.
[48] Marc H. Raibert,et al. Legged Robots That Balance , 1986, IEEE Expert.
[49] Peter J. Angeline,et al. An evolutionary algorithm that constructs recurrent neural networks , 1994, IEEE Trans. Neural Networks.
[50] Larry Rudolph,et al. Implementation Matters in Deep Policy Gradients: A Case Study on PPO and TRPO , 2020, ArXiv.
[51] Sangbae Kim,et al. The MIT super mini cheetah: A small, low-cost quadrupedal robot for dynamic locomotion , 2015, 2015 IEEE International Symposium on Safety, Security, and Rescue Robotics (SSRR).
[52] Larry S. Davis,et al. NISP: Pruning Networks Using Neuron Importance Score Propagation , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[53] Dmitry P. Vetrov,et al. Variational Dropout Sparsifies Deep Neural Networks , 2017, ICML.
[54] Atil Iscen,et al. Data Efficient Reinforcement Learning for Legged Robots , 2019, CoRL.