Learning to generate proactive and reactive behavior using a dynamic neural network model with time-varying variance prediction mechanism

This paper discusses a possible neurodynamic mechanism that enables self-organization of two basic behavioral modes, namely a ‘proactive mode’ and a ‘reactive mode,’ and of autonomous switching between these modes depending on the situation. In the proactive mode, actions are generated based on an internal prediction, whereas in the reactive mode actions are generated in response to sensory inputs in unpredictable situations. In order to investigate how these two behavioral modes can be self-organized and how autonomous switching between the two modes can be achieved, we conducted neurorobotics experiments by using our recently developed dynamic neural network model that has a capability to learn to predict time-varying variance of the observable variables. In a set of robot experiments under various conditions, the robot was required to imitate other’s movements consisting of alternating predictable and unpredictable patterns. The experimental results showed that the robot controlled by the neural network model was able to proactively imitate predictable patterns and reactively follow unpredictable patterns by autonomously switching its behavioral modes. Our analysis revealed that variance prediction mechanism can lead to self-organization of these abilities with sufficient robustness and generalization capabilities. Graphical Abstract

[1]  C. Breazeal,et al.  Robots that imitate humans , 2002, Trends in Cognitive Sciences.

[2]  Martin V. Butz,et al.  Anticipations, Brains, Individual and Social Behavior: An Introduction to Anticipatory Systems , 2007, SAB ABiALS.

[3]  R. Johansson,et al.  Eye–Hand Coordination in Object Manipulation , 2001, The Journal of Neuroscience.

[4]  Jun Tani,et al.  On-line Imitative Interaction with a Humanoid Robot Using a Dynamic Neural Network Model of a Mirror System , 2004, Adapt. Behav..

[5]  Karl J. Friston The free-energy principle: a rough guide to the brain? , 2009, Trends in Cognitive Sciences.

[6]  Karl J. Friston,et al.  Action understanding and active inference , 2011, Biological Cybernetics.

[7]  Jeffrey L. Elman,et al.  Finding Structure in Time , 1990, Cogn. Sci..

[8]  Geoffrey E. Hinton,et al.  Learning representations by back-propagating errors , 1986, Nature.

[9]  Stefan Schaal,et al.  Is imitation learning the route to humanoid robots? , 1999, Trends in Cognitive Sciences.

[10]  J. Mazziotta,et al.  Neural mechanisms of empathy in humans: A relay from neural systems for imitation to limbic areas , 2003, Proceedings of the National Academy of Sciences of the United States of America.

[11]  Hiroaki Arie,et al.  Synthetic Approach to Understanding Meta-level Cognition of Predictability in Generating Cooperative Behavior , 2013 .

[12]  Jun Tani,et al.  Dynamic and interactive generation of object handling behaviors by a small humanoid robot using a dynamic neural network model , 2006, Neural Networks.

[13]  C. Hofsten,et al.  Infants predict other people's action goals , 2006, Nature Neuroscience.

[14]  Aude Billard,et al.  Statistical Learning by Imitation of Competing Constraints in Joint Space and Task Space , 2009, Adv. Robotics.

[15]  Gerben G. Meyer,et al.  ANTICIPATORY BEHAVIOR IN ADAPTIVE LEARNING SYSTEMS: FROM BRAINS TO INDIVIDUAL AND SOCIAL BEHAVIOR , 2007 .

[16]  R W Pew,et al.  Levels of analysis in motor control. , 1974, Brain research.

[17]  Jun Tani,et al.  Model-based learning for mobile robot navigation from the dynamical systems perspective , 1996, IEEE Trans. Syst. Man Cybern. Part B.

[18]  D M Wolpert,et al.  Multiple paired forward and inverse models for motor control , 1998, Neural Networks.

[19]  Michael I. Jordan,et al.  Forward Models: Supervised Learning with a Distal Teacher , 1992, Cogn. Sci..

[20]  A. Meltzoff,et al.  Imitation of Facial and Manual Gestures by Human Neonates , 1977, Science.

[21]  Shigeki Sugano,et al.  Learning to Reproduce Fluctuating Time Series by Inferring Their Time-Dependent Stochastic Properties: Application in Robot Learning Via Tutoring , 2013, IEEE Transactions on Autonomous Mental Development.

[22]  Jun Tani,et al.  Emergence of Functional Hierarchy in a Multiple Timescale Neural Network Model: A Humanoid Robot Experiment , 2008, PLoS Comput. Biol..

[23]  J. Mazziotta,et al.  Cortical mechanisms of human imitation. , 1999, Science.

[24]  Scott T. Grafton,et al.  Motor Learning of Compatible and Incompatible Visuomotor Maps , 2001, Journal of Cognitive Neuroscience.

[25]  Aude Billard,et al.  On Learning, Representing, and Generalizing a Task in a Humanoid Robot , 2007, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[26]  J. Piaget Play, dreams and imitation in childhood , 1951 .

[27]  Yoko Yamaguchi,et al.  Advances in Cognitive Neurodynamics (III) , 2013, Springer Netherlands.

[28]  R. Johansson,et al.  Action plans used in action observation , 2003, Nature.

[29]  Jun Tani,et al.  Learning to generate articulated behavior through the bottom-up and the top-down interaction processes , 2003, Neural Networks.

[30]  T. Braver The variable nature of cognitive control: a dual mechanisms framework , 2012, Trends in Cognitive Sciences.

[31]  Ronald J. Williams,et al.  A Learning Algorithm for Continually Running Fully Recurrent Neural Networks , 1989, Neural Computation.