Grounding abstraction in sensory experience

In order to make appropriate decisions, intelligent creatures narrow their sensory information down to high-level, abstract knowledge. Inspired by recent findings in neuroscience on the role of mirror neurons in action-based abstraction, in this study we propose a two-phase framework whereby a reinforcement learning (RL) agent attempts to understand its environment via meaningful temporally extended concepts in an unsupervised way. Throughout the process of concept extraction, the reinforcement learning agent makes use of a model of short-term and long-term memory to retrieve meaningful concepts from its environment. Empirical results by using e-puck robots reveal the capability of the proposed approach in retrieving meaningful concepts from environments. Moreover, simulation results show that the proposed approach can help an agent learn once and apply its knowledge in other environments which have similar structure without any further learning.

[1]  Ben J. A. Kröse,et al.  Learning from delayed rewards , 1995, Robotics Auton. Syst..

[2]  Doina Precup,et al.  Using Rewards for Belief State Updates in Partially Observable Markov Decision Processes , 2005, ECML.

[3]  Doina Precup,et al.  Temporal abstraction in reinforcement learning , 2000, ICML 2000.

[4]  Richard S. Sutton,et al.  Temporal-Difference Networks with History , 2005, IJCAI.

[5]  Jonathan Schaeffer,et al.  Learning Partial-Order Macros from Solutions , 2005, ICAPS.

[6]  Hossein Mobahi,et al.  Concept Oriented Imitation Towards Verbal Human-Robot Interaction , 2005, Proceedings of the 2005 IEEE International Conference on Robotics and Automation.

[7]  Richard S. Sutton,et al.  Using Predictive Representations to Improve Generalization in Reinforcement Learning , 2005, IJCAI.

[8]  Doina Precup,et al.  Between MDPs and Semi-MDPs: A Framework for Temporal Abstraction in Reinforcement Learning , 1999, Artif. Intell..

[9]  Richard S. Sutton,et al.  Temporal-Difference Networks , 2004, NIPS.

[10]  G. Lakoff,et al.  The Brain's concepts: the role of the Sensory-motor system in conceptual knowledge , 2005, Cognitive neuropsychology.

[11]  Peter Stone,et al.  State Abstraction Discovery from Irrelevant State Variables , 2005, IJCAI.

[12]  Thomas J. Walsh,et al.  Towards a Unified Theory of State Abstraction for MDPs , 2006, AI&M.

[13]  G. Rizzolatti,et al.  Motor and cognitive functions of the ventral premotor cortex , 2002, Current Opinion in Neurobiology.

[14]  T. Zentall,et al.  Categorization, concept learning, and behavior analysis: an introduction. , 2002, Journal of the experimental analysis of behavior.

[15]  G. Rizzolatti,et al.  Parietal Lobe: From Action Organization to Intention Understanding , 2005, Science.

[16]  Andrew G. Barto,et al.  PolicyBlocks: An Algorithm for Creating Useful Macro-Actions in Reinforcement Learning , 2002, ICML.

[17]  Richard S. Sutton,et al.  TD(λ) networks: temporal-difference networks with eligibility traces , 2005, ICML.