暂无分享,去创建一个
[1] Robert E. Schapire,et al. A Reduction from Apprenticeship Learning to Classification , 2010, NIPS.
[2] Erik Talvitie,et al. Self-Correcting Models for Model-Based Reinforcement Learning , 2016, AAAI.
[3] Lantao Yu,et al. SeqGAN: Sequence Generative Adversarial Nets with Policy Gradient , 2016, AAAI.
[4] Jan Peters,et al. Data-Efficient Generalization of Robot Skills with Contextual Policy Search , 2013, AAAI.
[5] Hal Daumé,et al. Frustratingly Easy Domain Adaptation , 2007, ACL.
[6] Timo Aila,et al. A Style-Based Generator Architecture for Generative Adversarial Networks , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[7] Mario Lucic,et al. Are GANs Created Equal? A Large-Scale Study , 2017, NeurIPS.
[8] Bogdan Raducanu,et al. Memory Replay GANs: Learning to Generate New Categories without Forgetting , 2018, NeurIPS.
[9] Herke van Hoof,et al. Addressing Function Approximation Error in Actor-Critic Methods , 2018, ICML.
[10] Fahad Shahbaz Khan,et al. MineGAN: Effective Knowledge Transfer From GANs to Target Domains With Few Images , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[11] C. Villani. Optimal Transport: Old and New , 2008 .
[12] Eduardo F. Morales,et al. An Introduction to Reinforcement Learning , 2011 .
[13] Marc Peter Deisenroth,et al. Data-Efficient Reinforcement Learning with Probabilistic Model Predictive Control , 2017, AISTATS.
[14] Yuandong Tian,et al. Algorithmic Framework for Model-based Deep Reinforcement Learning with Theoretical Guarantees , 2018, ICLR.
[15] Pieter Abbeel,et al. Benchmarking Model-Based Reinforcement Learning , 2019, ArXiv.
[16] Sergey Levine,et al. Trust Region Policy Optimization , 2015, ICML.
[17] Sergey Levine,et al. When to Trust Your Model: Model-Based Policy Optimization , 2019, NeurIPS.
[18] Sergey Levine,et al. Neural Network Dynamics for Model-Based Deep Reinforcement Learning with Model-Free Fine-Tuning , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).
[19] D. Wood. The Computation of Polylogarithms , 1992 .
[20] Yuichi Yoshida,et al. Spectral Normalization for Generative Adversarial Networks , 2018, ICLR.
[21] Pieter Abbeel,et al. Model-Ensemble Trust-Region Policy Optimization , 2018, ICLR.
[22] Michael H. Bowling,et al. Apprenticeship learning using linear programming , 2008, ICML '08.
[23] Stefano Ermon,et al. Generative Adversarial Imitation Learning , 2016, NIPS.
[24] Geoffrey J. Gordon,et al. A Reduction of Imitation Learning and Structured Prediction to No-Regret Online Learning , 2010, AISTATS.
[25] Lorenzo Rosasco,et al. Learning Probability Measures with respect to Optimal Transport Metrics , 2012, NIPS.
[26] Alison L Gibbs,et al. On Choosing and Bounding Probability Metrics , 2002, math/0209021.
[27] Carl E. Rasmussen,et al. Gaussian Processes for Data-Efficient Learning in Robotics and Control , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[28] Honglak Lee,et al. Sample-Efficient Reinforcement Learning with Stochastic Ensemble Value Expansion , 2018, NeurIPS.
[29] Stefan Schaal,et al. Learning from Demonstration , 1996, NIPS.
[30] Tamim Asfour,et al. Model-Based Reinforcement Learning via Meta-Policy Optimization , 2018, CoRL.
[31] Kavosh Asadi,et al. Combating the Compounding-Error Problem with a Multi-step Model , 2019, ArXiv.
[32] Raymond Y. K. Lau,et al. Least Squares Generative Adversarial Networks , 2016, 2017 IEEE International Conference on Computer Vision (ICCV).
[33] Aaron C. Courville,et al. Improved Training of Wasserstein GANs , 2017, NIPS.
[34] Han Zhang,et al. Self-Attention Generative Adversarial Networks , 2018, ICML.
[35] Yuval Tassa,et al. MuJoCo: A physics engine for model-based control , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.
[36] John Blitzer,et al. Domain Adaptation with Structural Correspondence Learning , 2006, EMNLP.
[37] Alec Radford,et al. Proximal Policy Optimization Algorithms , 2017, ArXiv.
[38] Bogdan Raducanu,et al. Transferring GANs: generating images from limited data , 2018, ECCV.
[39] Sergey Levine,et al. Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor , 2018, ICML.
[40] Sergey Levine,et al. Deep Reinforcement Learning in a Handful of Trials using Probabilistic Dynamics Models , 2018, NeurIPS.
[41] Xiaohua Zhai,et al. A Large-Scale Study on Regularization and Normalization in GANs , 2018, ICML.