An Enactive approach to autonomous agent and robot learning

A novel way to model an agent interacting with an environment is introduced, called an Enactive Markov Decision Process (EMDP). An EMDP keeps perception and action embedded within sensorimotor schemes rather than dissociated. Instead of seeking a goal associated with a reward, as in reinforcement learning, an EMDP agent is driven by two forms of self-motivation: successfully enacting sequences of interactions (autotelic motivation), and preferably enacting interactions that have predefined positive values (interactional motivation). An EMDP learning algorithm is presented. Results show that the agent develops a rudimentary form of self-programming, along with active perception as it learns to master the sensorimotor contingencies afforded by its coupling with the environment.

[1]  R. Sternberg,et al.  The Psychology of Intelligence , 2002 .

[2]  Leslie Pack Kaelbling,et al.  Planning and Acting in Partially Observable Stochastic Domains , 1998, Artif. Intell..

[3]  Luc Steels,et al.  The Autotelic Principle , 2003, Embodied Artificial Intelligence.

[4]  R A Brooks,et al.  New Approaches to Robotics , 1991, Science.

[5]  James B. Marshall,et al.  Demonstrating sensemaking emergence in artificial agents: A method and an example , 2013 .

[6]  Tracy Brown,et al.  The Embodied Mind: Cognitive Science and Human Experience , 2002, Cybern. Hum. Knowing.

[7]  Murray Shanahan,et al.  Embodiment and the inner lifeCognition and Consciousness in the Space of Possible Minds , 2010 .

[8]  James F. Glazebrook Embodiment and the inner life: Cognition and consciousness in the space of possible minds, M. Shanahan. Oxford University Press (2010) , 2014, Cogn. Syst. Res..

[9]  D. Smith Consciousness in action , 2004, Synthese.

[10]  T. Ziemke The Construction of ‘Reality’ in the Robot: Constructivist Perspectives on Situated Artificial Intelligence and Adaptive Robotics , 2001 .

[11]  Andrew McCallum,et al.  Learning to Use Selective Attention and Short-Term Memory in Sequential Tasks , 1996 .

[12]  Pierre-Yves Oudeyer,et al.  Intrinsic Motivation Systems for Autonomous Mental Development , 2007, IEEE Transactions on Evolutionary Computation.

[13]  A. Noë,et al.  A sensorimotor account of vision and visual consciousness. , 2001, The Behavioral and brain sciences.

[14]  C. Scheier,et al.  From perception to action: the right direction? , 1994, Proceedings of PerAc '94. From Perception to Action.

[15]  Volker Krüger,et al.  Tracking in object action space , 2013, Comput. Vis. Image Underst..

[16]  Tom Ziemke,et al.  Enactive artificial intelligence: Investigating the systemic organization of life and mind , 2009, Artif. Intell..

[17]  James B. Marshall,et al.  Interactional Motivation in artificial systems: Between extrinsic and intrinsic motivation , 2012, 2012 IEEE International Conference on Development and Learning and Epigenetic Robotics (ICDL).

[18]  Gary L. Drescher,et al.  Made-up minds - a constructivist approach to artificial intelligence , 1991 .

[19]  Maja J. Matarić,et al.  Learning to Use Selective Attention and Short-Term Memory in Sequential Tasks , 1996 .

[20]  C. L. Giles,et al.  Sequence Learning - Paradigms, Algorithms, and Applications , 2001 .

[21]  Francesco Mondada,et al.  The e-puck, a Robot Designed for Education in Engineering , 2009 .

[22]  Frank E. Ritter,et al.  An intrinsically-motivated schema mechanism to model and simulate emergent cognition , 2012, Cognitive Systems Research.