Infinite-dimensional reservoir computing

Reservoir computing approximation and generalization bounds are proved for a new concept class of input/output systems that extends the so-called generalized Barron functionals to a dynamic context. This new class is characterized by the readouts with a certain integral representation built on infinite-dimensional state-space systems. It is shown that this class is very rich and possesses useful features and universal approximation properties. The reservoir architectures used for the approximation and estimation of elements in the new class are randomly generated echo state networks with either linear or ReLU activation functions. Their readouts are built using randomly generated neural networks in which only the output layer is trained (extreme learning machines or random feature neural networks). The results in the paper yield a fully implementable recurrent neural network-based learning algorithm with provable convergence guarantees that do not suffer from the curse of dimensionality.

[1]  J. Ortega,et al.  Quantum reservoir computing in finite dimensions. , 2022, Physical review. E.

[2]  J. Ortega,et al.  Transport in reservoir computing , 2022, Physica D: Nonlinear Phenomena.

[3]  F. Benth,et al.  Neural networks in Fréchet spaces , 2022, Annals of Mathematics and Artificial Intelligence.

[4]  G. Livieri,et al.  Designing Universal Causal Deep Learning Models: The Case of Infinite-Dimensional Dynamical Systems from Stochastic Analysis , 2022, ArXiv.

[5]  Ariel Neufeld,et al.  Chaotic Hedging with Iterated Integrals and Neural Networks , 2022, ArXiv.

[6]  Christa Cuchiero,et al.  Universal approximation theorems for continuous functions of c\`adl\`ag paths and L\'evy-type signature models , 2022, 2208.02293.

[7]  Yue Wang,et al.  Neural Operator with Regularity Structure for Modeling Dynamics Driven by SPDEs , 2022, ArXiv.

[8]  N. Matni,et al.  Single Trajectory Nonparametric Learning of Nonlinear Dynamics , 2022, COLT.

[9]  G Manjunath,et al.  Embedding information onto a dynamical system , 2021, Nonlinearity.

[10]  E. Weinan,et al.  Representation formulas and pointwise properties for Barron functions , 2020, Calculus of Variations and Partial Differential Equations.

[11]  E. Weinan,et al.  Approximation and Optimization Theory for Linear Continuous-Time Recurrent Neural Networks , 2022, J. Mach. Learn. Res..

[12]  B. Acciaio,et al.  Metric Hypertransformers are Universal Adapted Maps , 2022, ArXiv.

[13]  C. Salvi,et al.  Neural Stochastic PDEs: Resolution-Invariant Learning of Continuous Spatiotemporal Dynamics , 2021, NeurIPS.

[14]  Nikola B. Kovachki,et al.  Neural Operator: Learning Maps Between Function Spaces , 2021, ArXiv.

[15]  Juan-Pablo Ortega,et al.  Learning strange attractors with reservoir systems , 2021, ArXiv.

[16]  B. Hunt,et al.  A Hybrid Approach to Atmospheric Modeling That Combines Machine Learning With a Physics‐Based Numerical Model , 2021, Journal of Advances in Modeling Earth Systems.

[17]  Lukas Gonon,et al.  Random feature neural networks learn Black-Scholes type PDEs without curse of dimensionality , 2021, ArXiv.

[18]  Kohei Nakajima,et al.  Learning Temporal Quantum Tomography , 2021, Physical review letters.

[19]  E. Ott,et al.  Using Data Assimilation to Train a Hybrid Forecast System that Combines Machine-Learning and Knowledge-Based Components , 2021, Chaos.

[20]  Lukas Gonon,et al.  Fading memory echo state networks are universal , 2020, Neural Networks.

[21]  Nikola B. Kovachki,et al.  Fourier Neural Operator for Parametric Partial Differential Equations , 2020, ICLR.

[22]  Juan-Pablo Ortega,et al.  Dimension reduction in recurrent networks by canonicalization , 2020, Journal of Geometric Mechanics.

[23]  E Weinan,et al.  Towards a Mathematical Understanding of Neural Network-Based Machine Learning: what we know and what we don't , 2020, CSIAM Transactions on Applied Mathematics.

[24]  Kohei Nakajima,et al.  Higher-Order Quantum Reservoir Computing , 2020, ArXiv.

[25]  Anastasis Kratsios,et al.  Non-Euclidean Universal Approximation , 2020, NeurIPS.

[26]  E Weinan,et al.  On the Banach spaces associated with multi-layer ReLU networks: Function representation, approximation theory and gradient descent dynamics , 2020, CSIAM Transactions on Applied Mathematics.

[27]  Lukas Gonon,et al.  Memory and forecasting capacities of nonlinear recurrent networks , 2020, Physica D: Nonlinear Phenomena.

[28]  H. Nurdin,et al.  Temporal Information Processing on Noisy Quantum Computers , 2020, Physical Review Applied.

[29]  Peter Tino,et al.  Dynamical Systems as Temporal Feature Spaces , 2019, J. Mach. Learn. Res..

[30]  Juan-Pablo Ortega,et al.  Reservoir Computing Universality With Stochastic Inputs , 2018, IEEE Transactions on Neural Networks and Learning Systems.

[31]  G. Manjunath Supplementary Material for “Stability and Memory-loss go hand-in-hand: Three Results in Dynamics & Computation” , 2020 .

[32]  Wu Lei A PRIORI ESTIMATES OF THE POPULATION RISK FOR TWO-LAYER NEURAL NETWORKS , 2020 .

[33]  Lukas Gonon,et al.  Risk bounds for reservoir computing , 2019, ArXiv.

[34]  Juan-Pablo Ortega,et al.  Differentiable reservoir computing , 2019, J. Mach. Learn. Res..

[35]  Jiayin Chen,et al.  Learning nonlinear input–output maps with dissipative quantum systems , 2019, Quantum Information Processing.

[36]  Franz J. Király,et al.  Kernels for sequentially ordered data , 2016, J. Mach. Learn. Res..

[37]  Peter Tiño,et al.  Asymptotic Fisher memory of randomized linear symmetric Echo State Networks , 2018, Neurocomputing.

[38]  Juan-Pablo Ortega,et al.  Echo state networks are universal , 2018, Neural Networks.

[39]  Edward Ott,et al.  Attractor reconstruction by machine learning. , 2018, Chaos.

[40]  Jaideep Pathak,et al.  Model-Free Prediction of Large Spatiotemporally Chaotic Systems from Data: A Reservoir Computing Approach. , 2018, Physical review letters.

[41]  Juan-Pablo Ortega,et al.  Universal discrete-time reservoir computers with stochastic inputs and linear readouts using non-homogeneous state-affine systems , 2017, J. Mach. Learn. Res..

[42]  Jaideep Pathak,et al.  Using machine learning to replicate chaotic attractors and calculate Lyapunov exponents from data. , 2017, Chaos.

[43]  JAKE BOUVRIE,et al.  Kernel Methods for the Approximation of Nonlinear Systems , 2011, SIAM J. Control. Optim..

[44]  Hyunjoong Kim,et al.  Functional Analysis I , 2017 .

[45]  Romain Couillet,et al.  The Asymptotic Performance of Linear Echo State Neural Networks , 2016, J. Mach. Learn. Res..

[46]  Gábor Lugosi,et al.  Concentration Inequalities - A Nonasymptotic Theory of Independence , 2013, Concentration Inequalities.

[47]  Herbert Jaeger,et al.  Echo State Property Linked to an Input: Exploring a Fundamental Characteristic of Recurrent Neural Networks , 2013, Neural Computation.

[48]  Stefan J. Kiebel,et al.  Re-visiting the echo state property , 2012, Neural Networks.

[49]  Jake V. Bouvrie,et al.  Kernel Methods for the Approximation of Some Key Quantities of Nonlinear Systems , 2012, 1204.0563.

[50]  Benjamin Schrauwen,et al.  Recurrent Kernel Machines: Computing with Infinite Echo State Networks , 2012, Neural Computation.

[51]  Benjamin Schrauwen,et al.  Memory in linear recurrent neural networks in continuous time , 2010, Neural Networks.

[52]  S. Barry Cooper,et al.  Computability In Context: Computation and Logic in the Real World , 2009 .

[53]  C. Villani Optimal Transport: Old and New , 2008 .

[54]  Benjamin Recht,et al.  Random Features for Large-Scale Kernel Machines , 2007, NIPS.

[55]  Chee Kheong Siew,et al.  Extreme learning machine: Theory and applications , 2006, Neurocomputing.

[56]  Harald Haas,et al.  Harnessing Nonlinearity: Predicting Chaotic Systems and Saving Energy in Wireless Communication , 2004, Science.

[57]  Henry Markram,et al.  Real-Time Computing Without Stable States: A New Framework for Neural Computation Based on Perturbations , 2002, Neural Computation.

[58]  Herbert Jaeger,et al.  The''echo state''approach to analysing and training recurrent neural networks , 2001 .

[59]  Maxwell B. Stinchcombe,et al.  Neural network approximation of continuous functionals and continuous functions on compactifications , 1999, Neural Networks.

[60]  Hong Chen,et al.  Approximation capability to functions of several variables, nonlinear functionals, and operators by radial basis function neural networks , 1993, IEEE Trans. Neural Networks.

[61]  Andrew R. Barron,et al.  Universal approximation bounds for superpositions of a sigmoidal function , 1993, IEEE Trans. Inf. Theory.

[62]  Michael B. Matthews,et al.  Approximating nonlinear fading-memory operators using neural network models , 1993 .

[63]  Michael B. Matthews,et al.  On the uniform approximation of nonlinear discrete-time fading-memory systems using neural network models , 1992 .

[64]  Leon O. Chua,et al.  Fading memory and the problem of approximating nonlinear operators with volterra series , 1985 .

[65]  G. M.,et al.  Partial Differential Equations I , 2023, Applied Mathematical Sciences.