Dynamical approximation by recurrent neural networks

Abstract We examine the approximating power of recurrent networks for dynamical systems through an unbounded number of iterations. It is shown that the natural family of recurrent neural networks with saturated linear transfer functions and synaptic weight matrices of rank 1 are essentially equivalent to feedforward neural networks with recurrent layers. Therefore, they inherit the universal approximation property of real-valued functions in one variable in a stronger sense, namely through an unbounded number of iterations and approximation guaranteed to be within O(1/n), with n neurons and possibly lateral synapses allowed in the hidden-layer. However, they are not as complex in their dynamical behavior as systems defined by Turing machines. It is further proved that every continuous dynamical system can be approximated through all iterations, by both finite analog and boolean networks, when one requires approximation of given arbitrary exact orbits of the (perhaps unknown) map. This result no longer holds when the orbits of the given map are only available as contaminated orbits of the approximant net due to the presence of random noise (e.g., due to digital truncations of analog activations). Neural nets can nonetheless approximate large families of continuous maps, including chaotic maps and maps sensitive to initial conditions. A precise characterization of what maps can be approximated fault-tolerantly by analog and discrete neural networks for unboundedly many iterations remains an open problem.

[1]  Kurt Hornik,et al.  Multilayer feedforward networks are universal approximators , 1989, Neural Networks.

[2]  Vladik Kreinovich,et al.  Estimates of the Number of Hidden Units and Variation with Respect to Half-Spaces , 1997, Neural Networks.

[3]  Allan Pinkus,et al.  Multilayer Feedforward Networks with a Non-Polynomial Activation Function Can Approximate Any Function , 1991, Neural Networks.

[4]  Cristopher Moore,et al.  Closed-for Analytic Maps in One and Two Dimensions can Simulate Universal Turing Machines , 1999, Theor. Comput. Sci..

[5]  Max H. Garzon,et al.  Observability of Neural Network Behavior , 1993, NIPS.

[6]  J. Yorke,et al.  Pseudo-orbit shadowing in the family of tent maps , 1988 .

[7]  Ken-ichi Funahashi,et al.  On the approximate realization of continuous mappings by neural networks , 1989, Neural Networks.

[8]  Max H. Garzon,et al.  Models of massive parallelism: analysis of cellular automata and neural networks , 1995 .

[9]  Andrew R. Barron,et al.  Universal approximation bounds for superpositions of a sigmoidal function , 1993, IEEE Trans. Inf. Theory.

[10]  Eduardo Sontag,et al.  Turing computability with neural nets , 1991 .

[11]  M. Garzon,et al.  Real Computation with Cellular Automata , 1993 .

[12]  L. Jones A Simple Lemma on Greedy Approximation in Hilbert Space and Convergence Rates for Projection Pursuit Regression and Neural Network Training , 1992 .

[13]  Moore,et al.  Unpredictability and undecidability in dynamical systems. , 1990, Physical review letters.

[14]  George Cybenko,et al.  Approximation by superpositions of a sigmoidal function , 1992, Math. Control. Signals Syst..

[15]  C. Moore,et al.  Closed-form Analytic Maps in One and Two DimensionsCan Simulate Turing Machines , 1996 .

[16]  Max H. Garzon,et al.  On Dynamical Properties of Neural Networks , 1991, Complex Syst..

[17]  Akihiko Morimoto Some Stabilities of Group Automorphisms , 1981 .

[18]  Kurt Hornik,et al.  Universal approximation of an unknown mapping and its derivatives using multilayer feedforward networks , 1990, Neural Networks.

[19]  Michel Cosnard,et al.  Computability with Low-Dimensional Dynamical Systems , 1994, Theor. Comput. Sci..