暂无分享,去创建一个
Jonathan Scholz | Yannick Schroecker | Mel Vecerík | Jonathan Scholz | Mel Vecerík | Yannick Schroecker
[1] Nando de Freitas,et al. Playing hard exploration games by watching YouTube , 2018, NeurIPS.
[2] J. Andrew Bagnell,et al. Efficient Reductions for Imitation Learning , 2010, AISTATS.
[3] Samy Bengio,et al. Density estimation using Real NVP , 2016, ICLR.
[4] Richard S. Sutton,et al. Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.
[5] Jan Peters,et al. Relative Entropy Inverse Reinforcement Learning , 2011, AISTATS.
[6] Junichiro Yoshimoto,et al. Derivatives of Logarithmic Stationary Distributions for Policy Gradient Reinforcement Learning , 2010, Neural Computation.
[7] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[8] Sergey Levine,et al. DeepMimic , 2018, ACM Trans. Graph..
[9] Sergey Levine,et al. Time-Contrastive Networks: Self-Supervised Learning from Video , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).
[10] Yishay Mansour,et al. Policy Gradient Methods for Reinforcement Learning with Function Approximation , 1999, NIPS.
[11] Simon Osindero,et al. Conditional Generative Adversarial Nets , 2014, ArXiv.
[12] John N. Tsitsiklis,et al. Analysis of temporal-difference learning with function approximation , 1996, NIPS 1996.
[13] Maya Cakmak,et al. Trajectories and keyframes for kinesthetic teaching: A human-robot interaction perspective , 2012, 2012 7th ACM/IEEE International Conference on Human-Robot Interaction (HRI).
[14] Iain Murray,et al. Masked Autoregressive Flow for Density Estimation , 2017, NIPS.
[15] Ashley D. Edwards,et al. Forward-Backward Reinforcement Learning , 2018, ArXiv.
[16] Andrew Y. Ng,et al. Pharmacokinetics of a novel formulation of ivermectin after administration to goats , 2000, ICML.
[17] Sergey Levine,et al. Guided Cost Learning: Deep Inverse Optimal Control via Policy Optimization , 2016, ICML.
[18] Koray Kavukcuoglu,et al. Pixel Recurrent Neural Networks , 2016, ICML.
[19] Yannick Schroecker,et al. State Aware Imitation Learning , 2017, NIPS.
[20] Sergey Levine,et al. Recall Traces: Backtracking Models for Efficient Reinforcement Learning , 2018, ICLR.
[21] Stefano Ermon,et al. Generative Adversarial Imitation Learning , 2016, NIPS.
[22] Andrea Lockerd Thomaz,et al. Robot Learning from Human Teachers , 2014, Robot Learning from Human Teachers.
[23] J. Peng,et al. Efficient Learning and Planning Within the Dyna Framework , 1993, IEEE International Conference on Neural Networks.
[24] Martha White,et al. Organizing Experience: a Deeper Look at Replay Mechanisms for Sample-Based Planning in Continuous State Domains , 2018, IJCAI.
[25] Dean Pomerleau,et al. ALVINN, an autonomous land vehicle in a neural network , 2015 .
[26] Jitendra Malik,et al. Zero-Shot Visual Imitation , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[27] Sergey Levine,et al. Learning Robust Rewards with Adversarial Inverse Reinforcement Learning , 2017, ICLR 2017.
[28] Anind K. Dey,et al. Maximum Entropy Inverse Reinforcement Learning , 2008, AAAI.
[29] Hugo Larochelle,et al. MADE: Masked Autoencoder for Distribution Estimation , 2015, ICML.
[30] Heiga Zen,et al. WaveNet: A Generative Model for Raw Audio , 2016, SSW.
[31] Martin A. Riedmiller,et al. Leveraging Demonstrations for Deep Reinforcement Learning on Robotics Problems with Sparse Rewards , 2017, ArXiv.
[32] Max Welling,et al. Auto-Encoding Variational Bayes , 2013, ICLR.
[33] Geoffrey J. Gordon,et al. A Reduction of Imitation Learning and Structured Prediction to No-Regret Online Learning , 2010, AISTATS.
[34] Nando de Freitas,et al. Reinforcement and Imitation Learning for Diverse Visuomotor Skills , 2018, Robotics: Science and Systems.
[35] Jonathon Shlens,et al. Conditional Image Synthesis with Auxiliary Classifier GANs , 2016, ICML.
[36] Andrea Lockerd Thomaz,et al. Directing Policy Search with Interactively Taught Via-Points , 2016, AAMAS.