Equivalence Between Wasserstein and Value-Aware Model-based Reinforcement Learning
暂无分享,去创建一个
Kavosh Asadi | Evan Cater | Dipendra Misra | Michael L. Littman | Dipendra Kumar Misra | M. Littman | Kavosh Asadi | Evan Cater
[1] Martial Hebert,et al. Improving Multi-Step Prediction of Learned Time Series Models , 2015, AAAI.
[2] Csaba Szepesvári,et al. A Generalized Reinforcement-Learning Model: Convergence and Applications , 1996, ICML.
[3] R. Bellman. A Markovian Decision Process , 1957 .
[4] Kavosh Asadi,et al. Lipschitz Continuity in Model-based Reinforcement Learning , 2018, ICML.
[5] Alex M. Andrew,et al. Reinforcement Learning: : An Introduction , 1998 .
[6] J. Andrew Bagnell,et al. Agnostic System Identification for Model-Based Reinforcement Learning , 2012, ICML.
[7] Andrew W. Moore,et al. Reinforcement Learning: A Survey , 1996, J. Artif. Intell. Res..
[8] Pieter Abbeel,et al. Using inaccurate models in reinforcement learning , 2006, ICML.
[9] Jeff G. Schneider,et al. Autonomous helicopter control using reinforcement learning policy search methods , 2001, Proceedings 2001 ICRA. IEEE International Conference on Robotics and Automation (Cat. No.01CH37164).
[10] Hossein Mobahi,et al. Learning with a Wasserstein Loss , 2015, NIPS.
[11] Dale Schuurmans,et al. Bridging the Gap Between Value and Policy Based Reinforcement Learning , 2017, NIPS.
[12] Richard S. Sutton,et al. Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.
[13] Vicenç Gómez,et al. A unified view of entropy-regularized Markov decision processes , 2017, ArXiv.
[14] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[15] Daniel Nikovski,et al. Value-Aware Loss Function for Model-based Reinforcement Learning , 2017, AISTATS.
[16] Alejandro Agostini,et al. Reinforcement Learning with a Gaussian mixture model , 2010, The 2010 International Joint Conference on Neural Networks (IJCNN).
[17] Kavosh Asadi,et al. An Alternative Softmax Operator for Reinforcement Learning , 2016, ICML.
[18] Marc G. Bellemare,et al. A Distributional Perspective on Reinforcement Learning , 2017, ICML.
[19] C. Villani. Optimal Transport: Old and New , 2008 .
[20] Léon Bottou,et al. Wasserstein Generative Adversarial Networks , 2017, ICML.
[21] Erik Talvitie,et al. Model Regularization for Stable Sample Rollouts , 2014, UAI.