Hierarchical self-organizing learning systems for embodied intelligence

In this work, a framework of designing embodied intelligence (EI), along with the essential elements and their design principles, is proposed. This work intends to deploy the following design principles. Firstly, hierarchical self-organizing learning systems in the form of network made of neurons are the essential elements for building machine intelligence. The supervised, unsupervised and reinforcement learning are all necessary aspects of learning and are studied for machine intelligence building. In supervised learning, an efficient learning method for hierarchical multi-layered network structure is proposed and studied. In addition, a quantitative measure is proposed to quantify overfitting of a network in a given learning problem to determine proper network structure or proper learning period. In unsupervised learning, a sparsely-connected hierarchical network is developed to build the neural representations effectively and efficiently for densely-coded sensory inputs, and to enable the memory with large memory capacity and great fault tolerance. Secondly, the memory-based intelligence is not only for passive information processing and pattern storage. One of the critical capabilities of intelligence is continuous and intentional learning. Therefore, a goal creation system (GCS), also as a type of hierarchical self-organizing learning system based on simple and uniform structure, is presented that acts as the trigger for the agent's goal creation, memory management, active interaction and goal-oriented learning. As a self-organizing structure, it is responsible for evaluating actions according to goals, stimulating the learning of useful associations and representations for sensory inputs and motor outputs. It enables the more powerful hierarchical reinforcement learning, finds the ontology among sensory objects, creates the needs, and affects the agent's attention and perception. Biologically inspired structural design concept and the framework of EI proposed in this dissertation create a promising direction in the field of EI. It enables the desired capabilities for an intelligent machine to have, including the efficient, continuous and intentional learning, large representative memory capacity, and goal-oriented perception and action. The hierarchical self-organizing learning systems include all ingredients necessary to develop intelligence, and to motivate a machine to act on its own in its environment. Having the framework defined and design principles prepared, the future work will be done more consistently.

[1]  John E. Laird,et al.  Soar-RL: integrating reinforcement learning with Soar , 2005, Cognitive Systems Research.

[2]  Shane Legg,et al.  A Formal Measure of Machine Intelligence , 2006, ArXiv.

[3]  Peter Dayan,et al.  Competition and Multiple Cause Models , 1995, Neural Comput..

[4]  John H. R. Maunsell,et al.  Hierarchical organization and functional streams in the visual cortex , 1983, Trends in Neurosciences.

[5]  Xiaohui Xie,et al.  Selectively Grouping Neurons in Recurrent Networks of Lateral Inhibition , 2002, Neural Computation.

[6]  Rolf Pfeifer,et al.  Understanding intelligence , 2020, Inequality by Design.

[7]  W. Greenough,et al.  Experience-driven brain plasticity: beyond the synapse. , 2004, Neuron glia biology.

[8]  Rodney A. Brooks,et al.  Flesh and Machines: How Robots Will Change Us , 2002 .

[9]  C. Lee Giles,et al.  What Size Neural Network Gives Optimal Generalization? Convergence Properties of Backpropagation , 1998 .

[10]  David Zipser,et al.  Feature Discovery by Competive Learning , 1986, Cogn. Sci..

[11]  C. Stevens An evolutionary scaling law for the primate visual system and its basis in cortical function , 2001, Nature.

[12]  Michio Sugeno,et al.  A fuzzy-logic-based approach to qualitative modeling , 1993, IEEE Trans. Fuzzy Syst..

[13]  Yue Li,et al.  Neural Network with Memory and Cognitive Functions , 2005, ICANN.

[14]  Sridhar Mahadevan,et al.  Hierarchical Memory-Based Reinforcement Learning , 2000, NIPS.

[15]  M M Mesulam,et al.  Large‐scale neurocognitive networks and distributed processing for attention, language, and memory , 1990, Annals of neurology.

[16]  Peter Földiák,et al.  SPARSE CODING IN THE PRIMATE CORTEX , 2002 .

[17]  Geoffrey E. Hinton,et al.  Generative models for discovering sparse distributed representations. , 1997, Philosophical transactions of the Royal Society of London. Series B, Biological sciences.

[18]  Jacek M. Zurada,et al.  Introduction to artificial neural systems , 1992 .

[19]  Marcus Frean,et al.  The Upstart Algorithm: A Method for Constructing and Training Feedforward Neural Networks , 1990, Neural Computation.

[20]  James T. Kwok,et al.  Constructive algorithms for structure learning in feedforward neural networks for regression problems , 1997, IEEE Trans. Neural Networks.

[21]  Rudy Setiono,et al.  Feedforward Neural Network Construction Using Cross Validation , 2001, Neural Computation.

[22]  Lin Padgham,et al.  Representation and Reasoning for Goals in BDI Agents , 2002, ACSC.

[23]  F ROSENBLATT,et al.  The perceptron: a probabilistic model for information storage and organization in the brain. , 1958, Psychological review.

[24]  Anand S. Rao,et al.  Modeling Rational Agents within a BDI-Architecture , 1997, KR.

[25]  Guang-Bin Huang,et al.  Learning capability and storage capacity of two-hidden-layer feedforward networks , 2003, IEEE Trans. Neural Networks.

[26]  Orly Yadid-Pecht,et al.  High-speed and high-precision current winner-take-all circuit , 2005, IEEE Transactions on Circuits and Systems II: Express Briefs.

[27]  G. Lorentz Approximation of Functions , 1966 .

[28]  Rodney A. Brooks,et al.  Intelligence Without Reason , 1991, IJCAI.

[29]  Philipp Slusallek,et al.  Introduction to real-time ray tracing , 2005, SIGGRAPH Courses.

[30]  M. Csíkszentmihályi Creativity: Flow and the Psychology of Discovery and Invention , 1996 .

[31]  A. K. Pujari,et al.  Data Mining Techniques , 2006 .

[32]  Elie Bienenstock,et al.  Neural Networks and the Bias/Variance Dilemma , 1992, Neural Computation.

[33]  Stephen I. Gallant,et al.  Perceptron-based learning algorithms , 1990, IEEE Trans. Neural Networks.

[34]  Kurt Hornik,et al.  Multilayer feedforward networks are universal approximators , 1989, Neural Networks.

[35]  Tai Sing Lee,et al.  Hierarchical Bayesian inference in the visual cortex. , 2003, Journal of the Optical Society of America. A, Optics, image science, and vision.

[36]  David J. Field,et al.  Sparse coding with an overcomplete basis set: A strategy employed by V1? , 1997, Vision Research.

[37]  B L McNaughton,et al.  Path Integration and Cognitive Mapping in a Continuous Attractor Neural Network Model , 1997, The Journal of Neuroscience.

[38]  David J. Field,et al.  What Is the Goal of Sensory Coding? , 1994, Neural Computation.

[39]  Janusz A. Starzyk,et al.  CMOS current mode winner-take-all circuit with both excitatory and inhibitory feedback , 1993 .

[40]  M. Goldberg,et al.  Neuronal Activity in the Lateral Intraparietal Area and Spatial Attention , 2003, Science.

[41]  Ah Chung Tsoi,et al.  Lessons in Neural Network Training: Overfitting May be Harder than Expected , 1997, AAAI/IAAI.

[42]  Terrence J Sejnowski,et al.  Communication in Neuronal Networks , 2003, Science.

[43]  Lawrence D. Jackel,et al.  Backpropagation Applied to Handwritten Zip Code Recognition , 1989, Neural Computation.

[44]  Stephen Grossberg,et al.  Linking Visual Cortical Development to Visual Perception , 2000 .

[45]  Richard S. Sutton,et al.  Neural networks for control , 1990 .

[46]  Klaus-Robert Müller,et al.  Asymptotic statistical theory of overtraining and cross-validation , 1997, IEEE Trans. Neural Networks.

[47]  Janusz A. Starzyk,et al.  Challenges of Embodied Intelligence , 2006 .

[48]  Martin Rehn,et al.  A network that uses few active neurones to code visual input predicts the diverse shapes of cortical receptive fields , 2007, Journal of Computational Neuroscience.

[49]  A. Sloman,et al.  A Study of Motive Processing and Attention , 1993 .

[50]  Juyang Weng,et al.  Sparse representation from a winner-take-all neural network , 2004, 2004 IEEE International Joint Conference on Neural Networks (IEEE Cat. No.04CH37541).

[51]  J. Marshall THE SKIN SENSES , 1969 .

[52]  H. Akaike A new look at the statistical model identification , 1974 .

[53]  H. Barlow,et al.  Single Units and Sensation: A Neuron Doctrine for Perceptual Psychology? , 1972, Perception.

[54]  Joseph V. Hajnal,et al.  Registration Methodology: Concepts and Algorithms , 2001 .

[55]  Terence D. Sanger,et al.  Optimal unsupervised learning in a single-layer linear feedforward neural network , 1989, Neural Networks.

[56]  E. Lehmann Testing Statistical Hypotheses , 1960 .

[57]  A. Newell Unified Theories of Cognition , 1990 .

[58]  Guang-Bin Huang,et al.  Upper bounds on the number of hidden neurons in feedforward networks with arbitrary bounded nonlinear activation functions , 1998, IEEE Trans. Neural Networks.

[59]  Haibo He,et al.  Anticipation-Based Temporal Sequences Learning in Hierarchical Structure , 2007, IEEE Transactions on Neural Networks.

[60]  Roland Baddeley,et al.  An efficient code in V1? , 1996, Nature.

[61]  Lutz Prechelt,et al.  Automatic early stopping using cross validation: quantifying the criteria , 1998, Neural Networks.

[62]  Ian H. Witten,et al.  An Adaptive Optimal Controller for Discrete-Time Markov Environments , 1977, Inf. Control..

[63]  Russell Reed,et al.  Pruning algorithms-a survey , 1993, IEEE Trans. Neural Networks.

[64]  David J. Field,et al.  Emergence of simple-cell receptive field properties by learning a sparse code for natural images , 1996, Nature.

[65]  Derek Long,et al.  Goal Creation in Motivated Agents , 1995, ECAI Workshop on Agent Theories, Architectures, and Languages.

[66]  Patrick Brézillon,et al.  Lecture Notes in Artificial Intelligence , 1999 .

[67]  S. Grossberg,et al.  Adaptive pattern classification and universal recoding: I. Parallel development and coding of neural feature detectors , 1976, Biological Cybernetics.

[68]  E. Niebur,et al.  Growth patterns in the developing brain detected by using continuum mechanical tensor maps , 2022 .

[69]  Richard Granger,et al.  A cortical model of winner-take-all competition via lateral inhibition , 1992, Neural Networks.

[70]  B. McNaughton,et al.  Hippocampal synaptic enhancement and information storage within a distributed memory system , 1987, Trends in Neurosciences.

[71]  S Ullman,et al.  Shifts in selective visual attention: towards the underlying neural circuitry. , 1985, Human neurobiology.

[72]  C. Geula,et al.  Architecture of connectivity within a cingulo-fronto-parietal neurocognitive network for directed attention. , 1993, Archives of neurology.

[73]  M. Tovée,et al.  Representational capacity of face coding in monkeys. , 1996, Cerebral cortex.

[74]  Junichi Murata,et al.  Introduction and control of subgoals in reinforcement learning , 2007, Artificial Intelligence and Applications.

[75]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[76]  E. Oja Simplified neuron model as a principal component analyzer , 1982, Journal of mathematical biology.

[77]  R. Peyron,et al.  Functional imaging of brain responses to pain. A review and meta-analysis (2000) , 2000, Neurophysiologie Clinique/Clinical Neurophysiology.

[78]  Shun-ichi Amari,et al.  Network information criterion-determining the number of hidden units for an artificial neural network model , 1994, IEEE Trans. Neural Networks.

[79]  David Soto,et al.  Spatial attention and object-based attention: a comparison within a single task , 2004, Vision Research.

[80]  S. Yantis,et al.  Cortical mechanisms of space-based and object-based attentional control , 2003, Current Opinion in Neurobiology.

[81]  D. O. Hebb,et al.  The organization of behavior , 1988 .

[82]  V. Mountcastle,et al.  Response properties of neurons of cat's somatic sensory cortex to peripheral stimuli. , 1957, Journal of neurophysiology.

[83]  Richard Bellman,et al.  Adaptive Control Processes: A Guided Tour , 1961, The Mathematical Gazette.

[84]  E. Rolls Functions of neuronal networks in the hippocampus and neocortex in memory , 1989 .

[85]  David D. Vogel,et al.  Sparsely Connected, Hebbian Networks with Strikingly Large Storage Capacities , 1997, Neural Networks.

[86]  Zhen Zhu,et al.  Optimized Approximation Algorithm in Neural Networks Without Overfitting , 2008, IEEE Transactions on Neural Networks.

[87]  Teuvo Kohonen,et al.  Self-Organization and Associative Memory , 1988 .

[88]  Michael P. Georgeff,et al.  Decision-Making in an Embedded Reasoning System , 1989, IJCAI.

[89]  Petri Koistinen,et al.  Using additive noise in back-propagation training , 1992, IEEE Trans. Neural Networks.

[90]  Ethem Alpaydin,et al.  GAL: Networks That Grow When They Learn and Shrink When They Forget , 1994, Int. J. Pattern Recognit. Artif. Intell..

[91]  David B. Fogel An information criterion for optimal neural network selection , 1991, IEEE Trans. Neural Networks.

[92]  Sridhar Mahadevan,et al.  Recent Advances in Hierarchical Reinforcement Learning , 2003, Discret. Event Dyn. Syst..

[93]  Leslie G. Ungerleider Two cortical visual systems , 1982 .

[94]  Amparo Alonso-Betanzos,et al.  Linear-least-squares initialization of multilayer perceptrons through backpropagation of the desired response , 2005, IEEE Transactions on Neural Networks.

[95]  Shin'ichi Tamura,et al.  Capabilities of a four-layered feedforward neural network: four layers versus three , 1997, IEEE Trans. Neural Networks.

[96]  Chen-Fang Chang,et al.  Observer-based air fuel ratio control , 1998 .

[97]  James L. McClelland,et al.  An interactive activation model of context effects in letter perception: I. An account of basic findings. , 1981 .

[98]  Chuan Wang,et al.  Training neural networks with additive noise in the desired signal , 1999, IEEE Trans. Neural Networks.

[99]  John E. Hummel,et al.  Distributed representations of structure: A theory of analogical access and mapping. , 1997 .

[100]  Russell L. Smith,et al.  Intelligent motion control with an artificial cerebellum , 1998 .

[101]  M. DeWeese,et al.  Binary Spiking in Auditory Cortex , 2003, The Journal of Neuroscience.

[102]  G. Mangun,et al.  The neural mechanisms of top-down attentional control , 2000, Nature Neuroscience.

[103]  Tong Heng Lee,et al.  Geometrical interpretation and architecture selection of MLP , 2005, IEEE Transactions on Neural Networks.

[104]  Anthony K. P. Jones,et al.  Pain processing during three levels of noxious stimulation produces differential patterns of central activity , 1997, Pain.

[105]  J. Rissanen Stochastic Complexity and Modeling , 1986 .

[106]  J. Gallant,et al.  Natural Stimulation of the Nonclassical Receptive Field Increases Information Transmission Efficiency in V1 , 2002, The Journal of Neuroscience.

[107]  H. Guterman,et al.  Knowledge extraction from artificial neural network models , 1997, 1997 IEEE International Conference on Systems, Man, and Cybernetics. Computational Cybernetics and Simulation.

[108]  P. Lennie The Cost of Cortical Computation , 2003, Current Biology.

[109]  Anand S. Rao,et al.  An Abstract Architecture for Rational Agents , 1992, KR.

[110]  Abdesselam Bouzerdoum,et al.  A Pyramidal Neural Network For Visual Pattern Recognition , 2007, IEEE Transactions on Neural Networks.

[111]  Eero P. Simoncelli,et al.  Natural image statistics and neural representation. , 2001, Annual review of neuroscience.

[112]  David D. Vogel,et al.  A neural network model of memory and higher cognitive functions. , 2005, International journal of psychophysiology : official journal of the International Organization of Psychophysiology.

[113]  Amy L. Lansky,et al.  Reactive Reasoning and Planning , 1987, AAAI.

[114]  Dimitris A. Pados,et al.  On overfitting, generalization, and randomly expanded training sets , 2000, IEEE Trans. Neural Networks Learn. Syst..

[115]  D. George,et al.  A hierarchical Bayesian model of invariant pattern recognition in the visual cortex , 2005, Proceedings. 2005 IEEE International Joint Conference on Neural Networks, 2005..

[116]  Bruno A. Olshausen,et al.  A new window on sound , 2002, Nature Neuroscience.

[117]  Jen-Chuen Hsieh,et al.  Activation of the hypothalamus characterizes the acupuncture stimulation at the analgesic point in human: a positron emission tomography study , 2001, Neuroscience Letters.

[118]  Laurent Itti,et al.  A Goal Oriented Attention Guidance Model , 2002, Biologically Motivated Computer Vision.

[119]  Stanley C. Ahalt,et al.  Competitive learning algorithms for vector quantization , 1990, Neural Networks.

[120]  D. George,et al.  Invariant Pattern Recognition using Bayesian Inference on Hierarchical Sequences , 2004 .

[121]  Michael S. Lewicki,et al.  Efficient coding of natural sounds , 2002, Nature Neuroscience.

[122]  Kay Chen Tan,et al.  Estimating the Number of Hidden Neurons in a Feedforward Network Using the Singular Value Decomposition , 2006, IEEE Trans. Neural Networks.

[123]  Michael Luck,et al.  Motivated Behavior for Goal Adoption , 1998, DAI.

[124]  G. Edelman,et al.  Solving Bernstein's problem: a proposal for the development of coordinated movement by selection. , 1993, Child development.

[125]  Richard S. Sutton,et al.  Neuronlike adaptive elements that can solve difficult learning control problems , 1983, IEEE Transactions on Systems, Man, and Cybernetics.

[126]  Simon Haykin,et al.  Neural Networks: A Comprehensive Foundation , 1998 .

[127]  Pierre-Yves Oudeyer,et al.  Intrinsic Motivation Systems for Autonomous Mental Development , 2007, IEEE Transactions on Evolutionary Computation.

[128]  G. Pagnoni,et al.  Does Anticipation of Pain Affect Cortical Nociceptive Systems? , 2002, The Journal of Neuroscience.

[129]  Peter Dayan,et al.  Theoretical Neuroscience: Computational and Mathematical Modeling of Neural Systems , 2001 .

[130]  Keiji Tanaka,et al.  Inferotemporal cortex and object vision. , 1996, Annual review of neuroscience.

[131]  R. Melzack Pain and the neuromatrix in the brain. , 2001, Journal of dental education.

[132]  Paul-Gerhard Plöger,et al.  Echo State Networks used for Motor Control , 2005, Proceedings of the 2005 IEEE International Conference on Robotics and Automation.

[133]  R. Brooks Flesh and machines , 2001 .

[134]  Geoffrey E. Hinton,et al.  Distributed Representations , 1986, The Philosophy of Artificial Intelligence.

[135]  J L Gallant,et al.  Sparse coding and decorrelation in primary visual cortex during natural vision. , 2000, Science.

[136]  K. Ishii,et al.  Optimization of parameters of echo state network and its application to underwater robot , 2004, SICE 2004 Annual Conference.

[137]  D. Hill,et al.  Medical image registration , 2001, Physics in medicine and biology.

[138]  Maja J. Matarić,et al.  Learning to Use Selective Attention and Short-Term Memory in Sequential Tasks , 1996 .

[139]  D. J. Felleman,et al.  Distributed hierarchical processing in the primate cerebral cortex. , 1991, Cerebral cortex.

[140]  S. Grossberg How does the cerebral cortex work? Learning, attention, and grouping by the laminar circuits of visual cortex. , 1999, Spatial vision.

[141]  Jennie Si,et al.  Online learning control by association and reinforcement. , 2001, IEEE transactions on neural networks.

[142]  Kevin Swingler,et al.  Applying neural networks - a practical guide , 1996 .

[143]  Pieter R. Roelfsema,et al.  Object-based attention in the primary visual cortex of the macaque monkey , 1998, Nature.

[144]  Bram Bakker,et al.  Hierarchical Reinforcement Learning Based on Subgoal Discovery and Subpolicy Specialization , 2003 .

[145]  Takashi Yoneyama,et al.  Specification of Training Sets and the Number of Hidden Neurons for Multilayer Perceptrons , 2001, Neural Computation.

[146]  B Conrad,et al.  Region‐specific encoding of sensory and affective components of pain in the human brain: A positron emission tomography correlation analysis , 1999, Annals of neurology.