Resonant spatiotemporal learning in large random recurrent networks

Abstract. Taking a global analogy with the structure of perceptual biological systems, we present a system composed of two layers of real-valued sigmoidal neurons. The primary layer receives stimulating spatiotemporal signals, and the secondary layer is a fully connected random recurrent network. This secondary layer spontaneously displays complex chaotic dynamics. All connections have a constant time delay. We use for our experiments a Hebbian (covariance) learning rule. This rule slowly modifies the weights under the influence of a periodic stimulus. The effect of learning is twofold: (i) it simplifies the secondary-layer dynamics, which eventually stabilizes to a periodic orbit; and (ii) it connects the secondary layer to the primary layer, and realizes a feedback from the secondary to the primary layer. This feedback signal is added to the incoming signal, and matches it (i.e., the secondary layer performs a one-step prediction of the forthcoming stimulus). After learning, a resonant behavior can be observed: the system resonates with familiar stimuli, which activates a feedback signal. In particular, this resonance allows the recognition and retrieval of partial signals, and dynamic maintenence of the memory of past stimuli. This resonance is highly sensitive to the temporal relationships and to the periodicity of the presented stimuli. When we present stimuli which do not match in time or space, the feedback remains silent. The number of different stimuli for which resonant behavior can be learned is analyzed. As with Hopfield networks, the capacity is proportional to the size of the second, recurrent layer. Moreover, the high capacity displayed allows the implementation of our model on real-time systems interacting with their environment. Such an implementation is reported in the case of a simple behavior-based recognition task on a mobile robot. Finally, we present some functional analogies with biological systems in terms of autonomy and dynamic binding, and present some hypotheses on the computational role of feedback connections.

[1]  W. Freeman,et al.  How brains make chaos in order to make sense of the world , 1987, Behavioral and Brain Sciences.

[2]  T. Sejnowski,et al.  Storing covariance with nonlinearly interacting neurons , 1977, Journal of mathematical biology.

[3]  Wulfram Gerstner,et al.  A biologically motivated and analytically soluble model of collective oscillations in the cortex , 1993, Biological Cybernetics.

[4]  Philippe Gaussier,et al.  The visual homing problem: An example of robotics/biology cross fertilization , 2000, Robotics Auton. Syst..

[5]  Yasuo Kuniyoshi,et al.  Neural learning of embodied interaction dynamics , 1998, Neural Networks.

[6]  F. Crick Function of the thalamic reticular complex: the searchlight hypothesis. , 1984, Proceedings of the National Academy of Sciences of the United States of America.

[7]  M. Samuelides,et al.  Large deviations and mean-field theory for asymmetric random recurrent neural networks , 2002 .

[8]  Gregor Schöner,et al.  Dynamics of behavior: Theory and applications for autonomous robot architectures , 1995, Robotics Auton. Syst..

[9]  Peter König,et al.  Stimulus-Dependent Assembly Formation of Oscillatory Responses: II. Desynchronization , 1991, Neural Computation.

[10]  Ch. von der Malsburg,et al.  A neural cocktail-party processor , 1986, Biological Cybernetics.

[11]  Stefano Nolfi,et al.  Learning to perceive the world as articulated: an approach for hierarchical learning in sensory-motor systems , 1998, Neural Networks.

[12]  Adam Prügel-Bennett,et al.  Learning Synfire Chains: Turning Noise into Signal , 1996, Int. J. Neural Syst..

[13]  J J Hopfield,et al.  Neural networks and physical systems with emergent collective computational abilities. , 1982, Proceedings of the National Academy of Sciences of the United States of America.

[14]  S. Thorpe,et al.  Speed of processing in the human visual system , 1996, Nature.

[15]  Ronald J. Williams,et al.  A Learning Algorithm for Continually Running Fully Recurrent Neural Networks , 1989, Neural Computation.

[16]  Stephen Grossberg,et al.  Visual cortical mechanisms of perceptual grouping: interacting layers, networks, columns, and maps , 2000, Neural Networks.

[17]  Peter König,et al.  Stimulus-Dependent Assembly Formation of Oscillatory Responses: I. Synchronization , 1991, Neural Computation.

[18]  Stephen Grossberg,et al.  A massively parallel architecture for a self-organizing neural pattern recognition machine , 1988, Comput. Vis. Graph. Image Process..

[19]  Adam Prügel-Bennett,et al.  Analysis of synfire chains , 1995 .

[20]  Jean-Arcady Meyer,et al.  Learning to Perceive the World as Articulated: An Approach for Hierarchical Learning in Sensory-Motor Systems , 1998 .

[21]  G. Laurent,et al.  Distinct Mechanisms for Synchronization and Temporal Patterning of Odor-Encoding Neural Assemblies , 1996, Science.

[22]  Philippe Gaussier,et al.  Taking inspiration from the Hippocampus can help solving robotics problems , 1999, ESANN.

[23]  Jeffrey L. Elman,et al.  Finding Structure in Time , 1990, Cogn. Sci..

[24]  Bruno Cessac,et al.  Self-organization and dynamics reduction in recurrent networks: stimulus presentation and learning , 1998, Neural Networks.

[25]  F. Varela,et al.  Perception's shadow: long-distance synchronization of human brain activity , 1999, Nature.

[26]  L. Spaanenburg On the Compositionality of Neural Networks. , 2001 .

[27]  Sommers,et al.  Chaos in random neural networks. , 1988, Physical review letters.

[28]  E. Vaadia,et al.  Spatiotemporal firing patterns in the frontal cortex of behaving monkeys. , 1993, Journal of neurophysiology.

[29]  B. Cessac Increase in Complexity in Random Neural Networks , 1995 .

[30]  S. Amari,et al.  Characteristics of Random Nets of Analog Neuron-Like Elements , 1972, IEEE Trans. Syst. Man Cybern..

[31]  J. L. van Hemmen,et al.  Hebbian learning reconsidered: Representation of static and dynamic objects in associative neural nets , 1989, Biological Cybernetics.

[32]  Olivier Pinaud,et al.  Mean-field Theory and Synchronization in Random Recurrent Neural Networks , 2004, Neural Processing Letters.

[33]  P. Gaussiera,et al.  The visual homing problem : An example of robotics / biology cross fertilization , 1999 .

[34]  B. McNaughton,et al.  Multistability of cognitive maps in the hippocampus of old rats , 1997, Nature.

[35]  J. M. Hupé,et al.  Cortical feedback improves discrimination between figure and background by V1, V2 and V3 neurons , 1998, Nature.

[36]  W. Singer,et al.  Stimulus-specific neuronal oscillations in orientation columns of cat visual cortex. , 1989, Proceedings of the National Academy of Sciences of the United States of America.

[37]  Philippe Gaussier,et al.  PerAc: A neural architecture to control artificial animals , 1995, Robotics Auton. Syst..

[38]  B. Cessac,et al.  CONTROL OF THE TRANSITION TO CHAOS IN NEURAL NETWORKS WITH RANDOM CONNECTIVITY , 1993 .

[39]  Jean-Pierre Nadal,et al.  Sparsely coded neural networks , 1998 .