暂无分享,去创建一个
Pushmeet Kohli | Philip H. S. Torr | Nicolas Usunier | Gabriel Synnaeve | Zeming Lin | Nantas Nardelli | Pushmeet Kohli | Zeming Lin | Nicolas Usunier | Nantas Nardelli | Gabriel Synnaeve
[1] Wolfram Burgard,et al. Neural SLAM: Learning to Explore with External Memory , 2017, 1706.09520.
[2] Pieter Abbeel,et al. Value Iteration Networks , 2016, NIPS.
[3] Rob Fergus,et al. MazeBase: A Sandbox for Learning from Games , 2015, ArXiv.
[4] Mark J. Warshawsky,et al. A Modern Approach , 2005 .
[5] Vijay Kumar,et al. Memory Augmented Control Networks , 2017, ICLR.
[6] Satinder Singh,et al. Value Prediction Network , 2017, NIPS.
[7] Dimitri P. Bertsekas,et al. Dynamic Programming and Optimal Control, Two Volume Set , 1995 .
[8] Martha White,et al. Unifying Task Specification in Reinforcement Learning , 2016, ICML.
[9] Nicolas Usunier,et al. Episodic Exploration for Deep Deterministic Policies: An Application to StarCraft Micromanagement Tasks , 2016, ArXiv.
[10] Eike Rehder,et al. Cooperative Motion Planning for Non-Holonomic Agents with Value Iteration Networks , 2017, ArXiv.
[11] Yishay Mansour,et al. Policy Gradient Methods for Reinforcement Learning with Function Approximation , 1999, NIPS.
[12] Allan Jabri,et al. Universal Planning Networks: Learning Generalizable Representations for Visuomotor Control , 2018, ICML.
[13] Nando de Freitas,et al. Sample Efficient Actor-Critic with Experience Replay , 2016, ICLR.
[14] Razvan Pascanu,et al. Vector-based navigation using grid-like representations in artificial agents , 2018, Nature.
[15] Rahul Sukthankar,et al. Cognitive Mapping and Planning for Visual Navigation , 2017, International Journal of Computer Vision.
[16] Shimon Whiteson,et al. TreeQN and ATreeC: Differentiable Tree-Structured Models for Deep Reinforcement Learning , 2017, ICLR.
[17] John N. Tsitsiklis,et al. Actor-Critic Algorithms , 1999, NIPS.
[18] Luca Antiga,et al. Automatic differentiation in PyTorch , 2017 .
[19] Joelle Pineau,et al. The Bottleneck Simulator: A Model-based Deep Reinforcement Learning Approach , 2018, J. Artif. Intell. Res..
[20] Shimon Whiteson,et al. Stabilising Experience Replay for Deep Multi-Agent Reinforcement Learning , 2017, ICML.
[21] Peter Dayan,et al. Q-learning , 1992, Machine Learning.
[22] Shimon Whiteson,et al. TreeQN and ATreeC: Differentiable Tree Planning for Deep Reinforcement Learning , 2017, ICLR 2018.
[23] Razvan Pascanu,et al. Imagination-Augmented Agents for Deep Reinforcement Learning , 2017, NIPS.
[24] Richard S. Sutton,et al. Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.
[25] Florian Richoux,et al. TorchCraft: a Library for Machine Learning Research on Real-Time Strategy Games , 2016, ArXiv.
[26] Dimitri P. Bertsekas,et al. Dynamic Programming and Optimal Control, Vol. II , 1976 .
[27] Tom Schaul,et al. The Predictron: End-To-End Learning and Planning , 2016, ICML.
[28] Philip H. S. Torr,et al. DESIRE: Distant Future Prediction in Dynamic Scenes with Interacting Agents , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[29] Jelena Kovacevic,et al. Generalized Value Iteration Networks: Life Beyond Lattices , 2017, AAAI.
[30] Razvan Pascanu,et al. Learning to Navigate in Complex Environments , 2016, ICLR.
[31] Philip H. S. Torr,et al. Playing Doom with SLAM-Augmented Deep Reinforcement Learning , 2016, ArXiv.
[32] Alejandro Bordallo,et al. Counterfactual reasoning about intent for interactive navigation in dynamic environments , 2015, 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).
[33] Pieter Abbeel,et al. Learning Generalized Reactive Policies using Deep Neural Networks , 2017, ICAPS.
[34] K. Upton,et al. A modern approach , 1995 .
[35] Andrew W. Moore,et al. Reinforcement Learning: A Survey , 1996, J. Artif. Intell. Res..