Development of object manipulation through self-exploratory visuomotor experience
暂无分享,去创建一个
[1] Csaba Szepesvári,et al. Exploration-exploitation tradeoff using variance estimates in multi-armed bandits , 2009, Theor. Comput. Sci..
[2] S. Hochreiter,et al. REINFORCEMENT DRIVEN INFORMATION ACQUISITION IN NONDETERMINISTIC ENVIRONMENTS , 1995 .
[3] L. Baum,et al. A Maximization Technique Occurring in the Statistical Analysis of Probabilistic Functions of Markov Chains , 1970 .
[4] Jürgen Schmidhuber,et al. A possibility for implementing curiosity and boredom in model-building neural controllers , 1991 .
[5] Pierre-Yves Oudeyer,et al. Robust intrinsically motivated exploration and active learning , 2009, 2009 IEEE 8th International Conference on Development and Learning.
[6] B. Kuipers,et al. Learning to predict the effects of actions: Synergy between rules and landmarks , 2007, 2007 IEEE 6th International Conference on Development and Learning.
[7] Benjamin Kuipers,et al. Towards the Application of Reinforcement Learning to Undirected Developmental Learning , 2007 .
[8] Yiannis Demiris,et al. Learning Forward Models for Robots , 2005, IJCAI.
[9] Kohtaro Sabe,et al. A generative model for developmental understanding of visuomotor experience , 2011, 2011 IEEE International Conference on Development and Learning (ICDL).
[10] Kohtaro Sabe,et al. Reward-free learning using sparsely-connected hidden Markov models and local controllers , 2009, EpiRob.
[11] Peter Auer,et al. Finite-time Analysis of the Multiarmed Bandit Problem , 2002, Machine Learning.
[12] Richard S. Sutton,et al. Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.
[13] Kohtaro Sabe,et al. Self-regulation mechanism for continual autonomous learning in open-ended environments , 2009, EpiRob.
[14] Pierre-Yves Oudeyer,et al. The Playground Experiment: Task-Independent Development of a Curious Robot , 2005 .