暂无分享,去创建一个
Wilka Carvalho | Sungryull Sohn | Honglak Lee | Christopher Hoang | Jongwook Choi | Jongwook Choi | Sungryull Sohn | Honglak Lee | Wilka Carvalho | Christopher Hoang
[1] Kate Saenko,et al. Hierarchical Reinforcement Learning with Hindsight , 2018, ArXiv.
[2] Sergey Levine,et al. Temporal Difference Models: Model-Free Deep RL for Model-Based Control , 2018, ICLR.
[3] Leslie Pack Kaelbling,et al. Hierarchical Learning in Stochastic Domains: Preliminary Results , 1993, ICML.
[4] Leslie Pack Kaelbling,et al. Learning to Achieve Goals , 1993, IJCAI.
[5] Sergey Levine,et al. Visual Reinforcement Learning with Imagined Goals , 2018, NeurIPS.
[6] Ruslan Salakhutdinov,et al. Neural Topological SLAM for Visual Navigation , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[7] Wolfram Burgard,et al. Scheduled Intrinsic Drive: A Hierarchical Take on Intrinsically Motivated Exploration , 2019, ArXiv.
[8] Rob Fergus,et al. Composable Planning with Attributes , 2018, ICML.
[9] Yuval Tassa,et al. Continuous control with deep reinforcement learning , 2015, ICLR.
[10] Andrew W. Moore,et al. Multi-Value-Functions: Efficient Automatic Action Hierarchies for Multiple Goal MDPs , 1999, IJCAI.
[11] Peter Dayan,et al. Improving Generalization for Temporal Difference Learning: The Successor Representation , 1993, Neural Computation.
[12] Marek Wydmuch,et al. ViZDoom Competitions: Playing Doom From Pixels , 2018, IEEE Transactions on Games.
[13] Pieter Abbeel,et al. rlpyt: A Research Code Base for Deep Reinforcement Learning in PyTorch , 2019, ArXiv.
[14] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[15] Balaraman Ravindran,et al. Successor Options: An Option Discovery Framework for Reinforcement Learning , 2019, IJCAI.
[16] Marlos C. Machado,et al. Count-Based Exploration with the Successor Representation , 2018, AAAI.
[17] Sergey Levine,et al. Learning Actionable Representations with Goal-Conditioned Policies , 2018, ICLR.
[18] Vladlen Koltun,et al. Semi-parametric Topological Memory for Navigation , 2018, ICLR.
[19] Marcin Andrychowicz,et al. Hindsight Experience Replay , 2017, NIPS.
[20] Pieter Abbeel,et al. Sparse Graphical Memory for Robust Planning , 2020, NeurIPS.
[21] Sergey Levine,et al. Time-Contrastive Networks: Self-Supervised Learning from Video , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).
[22] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[23] Doina Precup,et al. Self-supervised Learning of Distance Functions for Goal-Conditioned Reinforcement Learning , 2019, ArXiv.
[24] Patrick M. Pilarski,et al. Horde: a scalable real-time architecture for learning knowledge from unsupervised sensorimotor interaction , 2011, AAMAS.
[25] Pieter Abbeel,et al. Hallucinative Topological Memory for Zero-Shot Visual Planning , 2020, ICML.
[26] Tom Schaul,et al. Universal Value Function Approximators , 2015, ICML.
[27] Tom Schaul,et al. Universal Successor Features Approximators , 2018, ICLR.
[28] Marc Pollefeys,et al. Episodic Curiosity through Reachability , 2018, ICLR.
[29] Hao Su,et al. Mapping State Space using Landmarks for Universal Goal Reaching , 2019, NeurIPS.
[30] Samuel Gershman,et al. Deep Successor Reinforcement Learning , 2016, ArXiv.
[31] Sergey Levine,et al. Search on the Replay Buffer: Bridging Planning and Reinforcement Learning , 2019, NeurIPS.
[32] Doina Precup,et al. Between MDPs and Semi-MDPs: A Framework for Temporal Abstraction in Reinforcement Learning , 1999, Artif. Intell..
[33] Yann LeCun,et al. Learning a similarity metric discriminatively, with application to face verification , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05).
[34] Tom Schaul,et al. Successor Features for Transfer in Reinforcement Learning , 2016, NIPS.
[35] Sergey Levine,et al. Skew-Fit: State-Covering Self-Supervised Reinforcement Learning , 2019, ICML.
[36] Marlos C. Machado,et al. Eigenoption Discovery through the Deep Successor Representation , 2017, ICLR.