暂无分享,去创建一个
[1] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[2] Sergey Levine,et al. Data-Efficient Hierarchical Reinforcement Learning , 2018, NeurIPS.
[3] Sergey Levine,et al. Model-Based Reinforcement Learning for Atari , 2019, ICLR.
[4] Zoubin Ghahramani,et al. Unsupervised learning of sensory-motor primitives , 2003, Proceedings of the 25th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (IEEE Cat. No.03CH37439).
[5] Pieter Abbeel,et al. Stochastic Neural Networks for Hierarchical Reinforcement Learning , 2016, ICLR.
[6] Pieter Abbeel,et al. Value Iteration Networks , 2016, NIPS.
[7] Han-Lim Choi,et al. Approximate Inference-Based Motion Planning by Learning and Exploiting Low-Dimensional Latent Variable Models , 2018, IEEE Robotics and Automation Letters.
[8] Sergey Levine,et al. Latent Space Policies for Hierarchical Reinforcement Learning , 2018, ICML.
[9] Emanuel Todorov,et al. General duality between optimal control and estimation , 2008, 2008 47th IEEE Conference on Decision and Control.
[10] Honglak Lee,et al. Learning Structured Output Representation using Deep Conditional Generative Models , 2015, NIPS.
[11] Sergey Levine,et al. Reinforcement Learning and Control as Probabilistic Inference: Tutorial and Review , 2018, ArXiv.
[12] Marc Toussaint,et al. Robot trajectory optimization using approximate inference , 2009, ICML '09.
[13] Carl E. Rasmussen,et al. Gaussian Processes for Data-Efficient Learning in Robotics and Control , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[14] Sergey Levine,et al. Dynamics-Aware Unsupervised Discovery of Skills , 2019, ICLR.
[15] Marc Toussaint,et al. On Stochastic Optimal Control and Reinforcement Learning by Approximate Inference , 2012, Robotics: Science and Systems.
[16] Sergey Levine,et al. Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor , 2018, ICML.
[17] Ole Winther,et al. A Disentangled Recognition and Nonlinear Dynamics Model for Unsupervised Learning , 2017, NIPS.
[18] Sergey Levine,et al. Deep Reinforcement Learning in a Handful of Trials using Probabilistic Dynamics Models , 2018, NeurIPS.
[19] Ruben Villegas,et al. Learning Latent Dynamics for Planning from Pixels , 2018, ICML.
[20] Nolan Wagener,et al. Information theoretic MPC for model-based reinforcement learning , 2017, 2017 IEEE International Conference on Robotics and Automation (ICRA).
[21] Luca Rigazio,et al. Path Integral Networks: End-to-End Differentiable Optimal Control , 2017, ArXiv.
[22] E. Todorov. Optimality principles in sensorimotor control , 2004, Nature Neuroscience.
[23] Sergey Levine,et al. Self-Consistent Trajectory Autoencoder: Hierarchical Reinforcement Learning with Trajectory Embeddings , 2018, ICML.
[24] Maximilian Karl,et al. Deep Variational Bayes Filters: Unsupervised Learning of State Space Models from Raw Data , 2016, ICLR.
[25] Yoshua Bengio,et al. Probabilistic Planning with Sequential Monte Carlo methods , 2018, ICLR.
[26] Nicolas Heess,et al. Hierarchical visuomotor control of humanoids , 2018, ICLR.
[27] Yee Whye Teh,et al. Neural probabilistic motor primitives for humanoid control , 2018, ICLR.
[28] Uri Shalit,et al. Structured Inference Networks for Nonlinear State Space Models , 2016, AAAI.
[29] Karol Hausman,et al. Learning an Embedding Space for Transferable Robot Skills , 2018, ICLR.
[30] Demis Hassabis,et al. Mastering the game of Go without human knowledge , 2017, Nature.
[31] Sergey Levine,et al. MCP: Learning Composable Hierarchical Control with Multiplicative Compositional Policies , 2019, NeurIPS.
[32] Jürgen Schmidhuber,et al. Recurrent World Models Facilitate Policy Evolution , 2018, NeurIPS.
[33] Sergey Levine,et al. Diversity is All You Need: Learning Skills without a Reward Function , 2018, ICLR.
[34] Joseph J. Lim,et al. Composing Complex Skills by Learning Transition Policies , 2018, ICLR.
[35] Sergey Levine,et al. DeepMimic , 2018, ACM Trans. Graph..
[36] Hilbert J. Kappen,et al. Adaptive Importance Sampling for Control and Inference , 2015, ArXiv.
[37] Alec Radford,et al. Proximal Policy Optimization Algorithms , 2017, ArXiv.
[38] Sergey Levine,et al. Meta-Reinforcement Learning of Structured Exploration Strategies , 2018, NeurIPS.
[39] Byron Boots,et al. Differentiable MPC for End-to-end Planning and Control , 2018, NeurIPS.
[40] Han-Lim Choi,et al. Adaptive path-integral autoencoder: representation learning and planning for dynamical systems , 2018, NeurIPS.
[41] Anind K. Dey,et al. Maximum Entropy Inverse Reinforcement Learning , 2008, AAAI.
[42] Steven M. LaValle,et al. Planning algorithms , 2006 .
[43] Max Welling,et al. Auto-Encoding Variational Bayes , 2013, ICLR.
[44] Byron Boots,et al. Continuous-time Gaussian process motion planning via probabilistic inference , 2017, Int. J. Robotics Res..