Composing Complex Skills by Learning Transition Policies
暂无分享,去创建一个
Joseph J. Lim | Sriram Somasundaram | Shao-Hua Sun | Youngwoon Lee | Edward S Hu | Youngwoon Lee | Shao-Hua Sun | E. Hu | Sriram Somasundaram
[1] Joshua B. Tenenbaum,et al. Hierarchical Deep Reinforcement Learning: Integrating Temporal Abstraction and Intrinsic Motivation , 2016, NIPS.
[2] Nando de Freitas,et al. Robust Imitation of Diverse Behaviors , 2017, NIPS.
[3] Jan Peters,et al. Probabilistic inference for determining options in reinforcement learning , 2016, Machine Learning.
[4] Stefano Ermon,et al. Generative Adversarial Imitation Learning , 2016, NIPS.
[5] Sergey Levine,et al. Divide-and-Conquer Reinforcement Learning , 2017, ICLR.
[6] Stefan Schaal,et al. Learning and generalization of motor skills by learning from demonstration , 2009, 2009 IEEE International Conference on Robotics and Automation.
[7] Joseph J. Lim,et al. Universal Agent for Disentangling Environments and Tasks , 2018, ICLR.
[8] Tom Schaul,et al. Unifying Count-Based Exploration and Intrinsic Motivation , 2016, NIPS.
[9] Richard S. Sutton,et al. Temporal credit assignment in reinforcement learning , 1984 .
[10] Doina Precup,et al. The Option-Critic Architecture , 2016, AAAI.
[11] Murray Shanahan,et al. Feature Control as Intrinsic Motivation for Hierarchical Reinforcement Learning , 2017, IEEE Transactions on Neural Networks and Learning Systems.
[12] Yuval Tassa,et al. Learning human behaviors from motion capture by adversarial imitation , 2017, ArXiv.
[13] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[14] Kate Saenko,et al. Hierarchical Actor-Critic , 2017, ArXiv.
[15] Honglak Lee,et al. Zero-Shot Task Generalization with Multi-Task Deep Reinforcement Learning , 2017, ICML.
[16] Raymond Y. K. Lau,et al. Least Squares Generative Adversarial Networks , 2016, 2017 IEEE International Conference on Computer Vision (ICCV).
[17] Yuval Tassa,et al. Emergence of Locomotion Behaviours in Rich Environments , 2017, ArXiv.
[18] Marcus Hutter,et al. Count-Based Exploration in Feature Space for Reinforcement Learning , 2017, IJCAI.
[19] Christoph H. Lampert,et al. Movement templates for learning of hitting and batting , 2010, 2010 IEEE International Conference on Robotics and Automation.
[20] Pushmeet Kohli,et al. Learning to Understand Goal Specifications by Modelling Reward , 2018, ICLR.
[21] Dan Klein,et al. Neural Module Networks , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[22] Richard Socher,et al. Hierarchical and Interpretable Skill Acquisition in Multi-task Reinforcement Learning , 2017, ICLR.
[23] Alec Radford,et al. Proximal Policy Optimization Algorithms , 2017, ArXiv.
[24] Sergey Levine,et al. Self-Consistent Trajectory Autoencoder: Hierarchical Reinforcement Learning with Trajectory Embeddings , 2018, ICML.
[25] Silvio Savarese,et al. Neural Task Programming: Learning to Generalize Across Hierarchical Tasks , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).
[26] Doina Precup,et al. Between MDPs and Semi-MDPs: A Framework for Temporal Abstraction in Reinforcement Learning , 1999, Artif. Intell..
[27] Martin A. Riedmiller,et al. Learning by Playing - Solving Sparse Reward Tasks from Scratch , 2018, ICML.
[28] Tom Schaul,et al. FeUdal Networks for Hierarchical Reinforcement Learning , 2017, ICML.
[29] Yuval Tassa,et al. MuJoCo: A physics engine for model-based control , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.
[30] Alexei A. Efros,et al. Curiosity-Driven Exploration by Self-Supervised Prediction , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[31] Glen Berseth,et al. DeepLoco , 2017, ACM Trans. Graph..
[32] Nan Jiang,et al. Hierarchical Imitation and Reinforcement Learning , 2018, ICML.
[33] Ruofan Kong,et al. Deep Reinforcement Learning for Dexterous Manipulation with Concept Networks , 2017, ArXiv.
[34] Sergey Levine,et al. Trust Region Policy Optimization , 2015, ICML.
[35] Andrew Y. Ng,et al. Policy Invariance Under Reward Transformations: Theory and Application to Reward Shaping , 1999, ICML.
[36] Pieter Abbeel,et al. Meta Learning Shared Hierarchies , 2017, ICLR.
[37] Oliver Kroemer,et al. Learning to select and generalize striking movements in robot table tennis , 2012, AAAI Fall Symposium: Robots Learning Interactively from Human Teachers.
[38] Dan Klein,et al. Modular Multitask Reinforcement Learning with Policy Sketches , 2016, ICML.