WWN: Integration with coarse-to-fine, supervised and reinforcement learning

The cost of autonomous development is substantial. Although supervised learning is effective, the cost demand on teachers is often too high to be constantly applied. Reinforcement learning can take advantage of physical reality due to environmental feedback and inspections. Information required in reinforcement learning is not as specific as is required in supervised learning. Integration theories, methods, and analysis of these two learning strategies are still rare in the literature although such integration has been well known in the animal kingdom. Based on our prior work on a general purpose framework called Developmental Network and its embodiment Where-What-Network, we present our theory, method, and analysis for integration of supervised learning and reinforcement learning in this paper. Different from all other known work on reinforcement learning, this DN framework uses fully emergent representation to avoid the brittleness and task-specific representations. Central in the integration is not just to provide a freedom for the teacher to choose the mode of learning, which is necessary especially when the physical non-living world is an implicit teacher, but the mechanism of scaffolding. In our experiment the scaffolding is reflected by allowing the location motor(LM) neurons to gradually refine representation through splitting(mitosis) in a coarse to fine scheme. We report our experimental work in a very challenging learning setting: both object and backgrounds are unknown(cluttered settings) and concepts(e.g. location and type) emerge from agent-environment interactions, instead of rigidly handcrafted.

[1]  Xinjing Wang,et al.  A unified learning paradigm for large-scale personalized information management , 2005, Conference, Emerging Information Technology 2005..

[2]  Juyang Weng,et al.  Modeling dopamine and serotonin systems in a visual recognition network , 2011, The 2011 International Joint Conference on Neural Networks.

[3]  J. Elman Learning and development in neural networks: the importance of starting small , 1993, Cognition.

[4]  Juyang Weng,et al.  Dually Optimal Neuronal Layers: Lobe Component Analysis , 2009, IEEE Transactions on Autonomous Mental Development.

[5]  Juyang Weng,et al.  Skull-closed autonomous development: WWN-6 using natural video , 2012, The 2012 International Joint Conference on Neural Networks (IJCNN).

[6]  Juyang Weng,et al.  Where What Network 3 : Developmental Top-Down Attention with Multiple Meaningful Foregrounds , 2010 .

[7]  S. Wanska,et al.  The relationship of spatial concept development to the acquisition of locative understanding. , 1984, The Journal of genetic psychology.

[8]  Michael T. Rosenstein,et al.  Supervised Actor‐Critic Reinforcement Learning , 2012 .

[9]  Kui Qian,et al.  Modeling the effects of neuromodulation on internal brain areas: Serotonin and dopamine , 2013, The 2013 International Joint Conference on Neural Networks (IJCNN).

[10]  Judith A. Langer,et al.  Instructional Scaffolding: Reading and Writing as Natural Language Activities , 1983, Language Arts.

[11]  Juyang Weng,et al.  Developmental Stereo: Emergence of Disparity Preference in Models of the Visual Cortex , 2009, IEEE Transactions on Autonomous Mental Development.

[12]  Ron Sun,et al.  Cognition and Multi-Agent Interaction: The CLARION Cognitive Architecture: Extending Cognitive Modeling to Social Simulation , 2005 .

[13]  Juyang Weng,et al.  WWN-text: Cortex-like language acquisition with “what” and “where” , 2010, 2010 IEEE 9th International Conference on Development and Learning.

[14]  Juyang Weng,et al.  Where-What Network 3: Developmental top-down attention for multiple foregrounds and complex backgrounds , 2010, The 2010 International Joint Conference on Neural Networks (IJCNN).

[15]  Juyang Weng,et al.  Where-what network 1: “Where” and “what” assist each other through top-down connections , 2008, 2008 7th IEEE International Conference on Development and Learning.

[16]  Min Guo,et al.  Reinforcement Learning Neural Network to the Problem of Autonomous Mobile Robot Obstacle Avoidance , 2005, 2005 International Conference on Machine Learning and Cybernetics.

[17]  Juyang Weng,et al.  Where-What Network 5: Dealing with scales for objects in complex backgrounds , 2011, The 2011 International Joint Conference on Neural Networks.