One Demonstration Imitation Learning
暂无分享,去创建一个
[1] Matthew W. Hoffman,et al. Modular Meta-Learning with Shrinkage , 2019, NeurIPS.
[2] Anind K. Dey,et al. Maximum Entropy Inverse Reinforcement Learning , 2008, AAAI.
[3] Alexei A. Efros,et al. Curiosity-Driven Exploration by Self-Supervised Prediction , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[4] S. Hochreiter,et al. REINFORCEMENT DRIVEN INFORMATION ACQUISITION IN NONDETERMINISTIC ENVIRONMENTS , 1995 .
[5] Jürgen Schmidhuber,et al. A possibility for implementing curiosity and boredom in model-building neural controllers , 1991 .
[6] Filip De Turck,et al. VIME: Variational Information Maximizing Exploration , 2016, NIPS.
[7] Sebastian Thrun,et al. Is Learning The n-th Thing Any Easier Than Learning The First? , 1995, NIPS.
[8] Sergey Levine,et al. Unsupervised Learning via Meta-Learning , 2018, ICLR.
[9] Pieter Abbeel,et al. Learning Plannable Representations with Causal InfoGAN , 2018, NeurIPS.
[10] Jürgen Schmidhuber,et al. Reinforcement Learning with Self-Modifying Policies , 1998, Learning to Learn.
[11] Qiang Liu,et al. Learning to Explore via Meta-Policy Gradient , 2018, ICML.
[12] Andrew Y. Ng,et al. Near-Bayesian exploration in polynomial time , 2009, ICML '09.
[13] Marcin Andrychowicz,et al. Hindsight Experience Replay , 2017, NIPS.
[14] Sergey Levine,et al. Efficient Off-Policy Meta-Reinforcement Learning via Probabilistic Context Variables , 2019, ICML.
[15] Sergey Levine,et al. One-Shot Visual Imitation Learning via Meta-Learning , 2017, CoRL.
[16] Stefano Ermon,et al. Generative Adversarial Imitation Learning , 2016, NIPS.
[17] Sergey Levine,et al. Diversity is All You Need: Learning Skills without a Reward Function , 2018, ICLR.
[18] Marcin Andrychowicz,et al. One-Shot Imitation Learning , 2017, NIPS.
[19] Max Welling,et al. Auto-Encoding Variational Bayes , 2013, ICLR.
[20] P. Bromiley. Products and Convolutions of Gaussian Probability Density Functions , 2013 .
[21] Marijn F. Stollenga,et al. Continual curiosity-driven skill acquisition from high-dimensional video inputs for humanoid robots , 2017, Artif. Intell..
[22] Jitendra Malik,et al. Combining self-supervised learning and imitation for vision-based rope manipulation , 2017, 2017 IEEE International Conference on Robotics and Automation (ICRA).
[23] Philippe Beaudoin,et al. Independently Controllable Factors , 2017, ArXiv.
[24] Geoffrey J. Gordon,et al. A Reduction of Imitation Learning and Structured Prediction to No-Regret Online Learning , 2010, AISTATS.
[25] Sebastian Nowozin,et al. Meta-Learning Probabilistic Inference for Prediction , 2018, ICLR.
[26] Lorenzo Natale,et al. Learning latent state representation for speeding up exploration , 2019, ArXiv.
[27] Sergey Levine,et al. Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor , 2018, ICML.
[28] Anca D. Dragan,et al. Learning a Prior over Intent via Meta-Inverse Reinforcement Learning , 2018, ICML.
[29] Jitendra Malik,et al. Zero-Shot Visual Imitation , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[30] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[31] Allan Jabri,et al. Universal Planning Networks , 2018, ICML.
[32] Ken Goldberg,et al. Deep Imitation Learning for Complex Manipulation Tasks from Virtual Reality Teleoperation , 2017, ICRA.
[33] Sergey Levine,et al. Unsupervised Meta-Learning for Reinforcement Learning , 2018, ArXiv.
[34] Martin A. Riedmiller,et al. Embed to Control: A Locally Linear Latent Dynamics Model for Control from Raw Images , 2015, NIPS.
[35] Richard S. Zemel,et al. Prototypical Networks for Few-shot Learning , 2017, NIPS.
[36] Nando de Freitas,et al. Robust Imitation of Diverse Behaviors , 2017, NIPS.
[37] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[38] Koray Kavukcuoglu,et al. Neural scene representation and rendering , 2018, Science.
[39] Richard S. Zemel,et al. SMILe: Scalable Meta Inverse Reinforcement Learning through Context-Conditional Policies , 2019, NeurIPS.
[40] Alexandre Lacoste,et al. TADAM: Task dependent adaptive metric for improved few-shot learning , 2018, NeurIPS.
[41] Pieter Abbeel,et al. Apprenticeship learning via inverse reinforcement learning , 2004, ICML.
[42] Jürgen Schmidhuber,et al. Learning skills from play: Artificial curiosity on a Katana robot arm , 2012, The 2012 International Joint Conference on Neural Networks (IJCNN).