暂无分享,去创建一个
Mo Yu | Shiyu Chang | Xiaoxiao Guo | Gerald Tesauro | Murray Campbell | G. Tesauro | Xiaoxiao Guo | Murray Campbell | Mo Yu | Shiyu Chang
[1] Pieter Abbeel,et al. Third-Person Imitation Learning , 2017, ICLR.
[2] Nando de Freitas,et al. Reinforcement and Imitation Learning for Diverse Visuomotor Skills , 2018, Robotics: Science and Systems.
[3] Dean Pomerleau,et al. ALVINN, an autonomous land vehicle in a neural network , 2015 .
[4] Stefan Schaal,et al. Robot Programming by Demonstration , 2009, Springer Handbook of Robotics.
[5] Marcin Andrychowicz,et al. Overcoming Exploration in Reinforcement Learning with Demonstrations , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).
[6] Peter Stone,et al. Behavioral Cloning from Observation , 2018, IJCAI.
[7] Geoffrey J. Gordon,et al. A Reduction of Imitation Learning and Structured Prediction to No-Regret Online Learning , 2010, AISTATS.
[8] Richard S. Sutton,et al. Dyna, an integrated architecture for learning, planning, and reacting , 1990, SGAR.
[9] Zachary Chase Lipton,et al. Efficient Exploration for Dialogue Policy Learning with BBQ Networks & Replay Buffer Spiking , 2016 .
[10] Brett Browning,et al. A survey of robot learning from demonstration , 2009, Robotics Auton. Syst..
[11] Stefano Ermon,et al. Generative Adversarial Imitation Learning , 2016, NIPS.
[12] Marlos C. Machado,et al. Eigenoption Discovery through the Deep Successor Representation , 2017, ICLR.
[13] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[14] Nando de Freitas,et al. Playing hard exploration games by watching YouTube , 2018, NeurIPS.
[15] Stefan Schaal,et al. Is imitation learning the route to humanoid robots? , 1999, Trends in Cognitive Sciences.
[16] Olivier Pietquin,et al. Observational Learning by Reinforcement Learning , 2017, AAMAS.
[17] J. Andrew Bagnell,et al. Maximum margin planning , 2006, ICML.
[18] Wojciech Zaremba,et al. OpenAI Gym , 2016, ArXiv.
[19] Marcin Andrychowicz,et al. One-Shot Imitation Learning , 2017, NIPS.
[20] Eyal Amir,et al. Bayesian Inverse Reinforcement Learning , 2007, IJCAI.
[21] Bolei Zhou,et al. Visual Question Generation as Dual Task of Visual Question Answering , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[22] Andrea Lockerd Thomaz,et al. Exploration from Demonstration for Interactive Reinforcement Learning , 2016, AAMAS.
[23] Anind K. Dey,et al. Maximum Entropy Inverse Reinforcement Learning , 2008, AAAI.
[24] Sergey Levine,et al. One-Shot Imitation from Observing Humans via Domain-Adaptive Meta-Learning , 2018, Robotics: Science and Systems.
[25] Gaurav S. Sukhatme,et al. Combining Model-Based and Model-Free Updates for Trajectory-Centric Reinforcement Learning , 2017, ICML.
[26] Thomas G. Dietterich. The MAXQ Method for Hierarchical Reinforcement Learning , 1998, ICML.
[27] Tom Schaul,et al. Deep Q-learning From Demonstrations , 2017, AAAI.
[28] Traian Rebedea,et al. Playing Atari Games with Deep Reinforcement Learning and Human Checkpoint Replay , 2016, ArXiv.
[29] Sergey Levine,et al. Nonlinear Inverse Reinforcement Learning with Gaussian Processes , 2011, NIPS.
[30] Sergey Levine,et al. Imitation from Observation: Learning to Imitate Behaviors from Raw Video via Context Translation , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).
[31] Andrew Y. Ng,et al. Pharmacokinetics of a novel formulation of ivermectin after administration to goats , 2000, ICML.
[32] Sergey Levine,et al. Guided Cost Learning: Deep Inverse Optimal Control via Policy Optimization , 2016, ICML.
[33] Pieter Abbeel,et al. Apprenticeship learning via inverse reinforcement learning , 2004, ICML.
[34] Olivier Spanjaard,et al. Reducing the Number of Queries in Interactive Value Iteration , 2015, ADT.
[35] Jianfeng Gao,et al. BBQ-Networks: Efficient Exploration in Deep Reinforcement Learning for Task-Oriented Dialogue Systems , 2016, AAAI.
[36] Honglak Lee,et al. Action-Conditional Video Prediction using Deep Networks in Atari Games , 2015, NIPS.