Grounding abstraction in sensory experience
暂无分享,去创建一个
[1] Ben J. A. Kröse,et al. Learning from delayed rewards , 1995, Robotics Auton. Syst..
[2] Doina Precup,et al. Using Rewards for Belief State Updates in Partially Observable Markov Decision Processes , 2005, ECML.
[3] Doina Precup,et al. Temporal abstraction in reinforcement learning , 2000, ICML 2000.
[4] Richard S. Sutton,et al. Temporal-Difference Networks with History , 2005, IJCAI.
[5] Jonathan Schaeffer,et al. Learning Partial-Order Macros from Solutions , 2005, ICAPS.
[6] Hossein Mobahi,et al. Concept Oriented Imitation Towards Verbal Human-Robot Interaction , 2005, Proceedings of the 2005 IEEE International Conference on Robotics and Automation.
[7] Richard S. Sutton,et al. Using Predictive Representations to Improve Generalization in Reinforcement Learning , 2005, IJCAI.
[8] Doina Precup,et al. Between MDPs and Semi-MDPs: A Framework for Temporal Abstraction in Reinforcement Learning , 1999, Artif. Intell..
[9] Richard S. Sutton,et al. Temporal-Difference Networks , 2004, NIPS.
[10] G. Lakoff,et al. The Brain's concepts: the role of the Sensory-motor system in conceptual knowledge , 2005, Cognitive neuropsychology.
[11] Peter Stone,et al. State Abstraction Discovery from Irrelevant State Variables , 2005, IJCAI.
[12] Thomas J. Walsh,et al. Towards a Unified Theory of State Abstraction for MDPs , 2006, AI&M.
[13] G. Rizzolatti,et al. Motor and cognitive functions of the ventral premotor cortex , 2002, Current Opinion in Neurobiology.
[14] T. Zentall,et al. Categorization, concept learning, and behavior analysis: an introduction. , 2002, Journal of the experimental analysis of behavior.
[15] G. Rizzolatti,et al. Parietal Lobe: From Action Organization to Intention Understanding , 2005, Science.
[16] Andrew G. Barto,et al. PolicyBlocks: An Algorithm for Creating Useful Macro-Actions in Reinforcement Learning , 2002, ICML.
[17] Richard S. Sutton,et al. TD(λ) networks: temporal-difference networks with eligibility traces , 2005, ICML.