Automatic Discovery of Subgoals in Reinforcement Learning Using Unique-Dreiction Value
暂无分享,去创建一个
Zhongzhi Shi | Rui Huang | Chuan Shi | Zhongzhi Shi | Chuan Shi | Rui Huang
[1] Doina Precup,et al. Temporal abstraction in reinforcement learning , 2000, ICML 2000.
[2] Bruce L. Digney,et al. Learning hierarchical control structures for multiple tasks and changing environments , 1998 .
[3] Thomas G. Dietterich. Hierarchical Reinforcement Learning with the MAXQ Value Function Decomposition , 1999, J. Artif. Intell. Res..
[4] Doina Precup,et al. Between MDPs and Semi-MDPs: A Framework for Temporal Abstraction in Reinforcement Learning , 1999, Artif. Intell..
[5] Shie Mannor,et al. Dynamic abstraction in reinforcement learning via clustering , 2004, ICML.
[6] Magnus Borga,et al. Hierarchical Reinforcement Learning , 1993 .
[7] Doina Precup,et al. Learning Options in Reinforcement Learning , 2002, SARA.
[8] Ronald E. Parr,et al. Hierarchical control and learning for markov decision processes , 1998 .
[9] Carla E. Brodley,et al. Proceedings of the twenty-first international conference on Machine learning , 2004, International Conference on Machine Learning.
[10] Alicia P. Wolfe,et al. Identifying useful subgoals in reinforcement learning by local graph partitioning , 2005, ICML.
[11] Andrew G. Barto,et al. Automatic Discovery of Subgoals in Reinforcement Learning using Diverse Density , 2001, ICML.