Infinite-dimensional reservoir computing
暂无分享,去创建一个
[1] J. Ortega,et al. Quantum reservoir computing in finite dimensions. , 2022, Physical review. E.
[2] J. Ortega,et al. Transport in reservoir computing , 2022, Physica D: Nonlinear Phenomena.
[3] F. Benth,et al. Neural networks in Fréchet spaces , 2022, Annals of Mathematics and Artificial Intelligence.
[4] G. Livieri,et al. Designing Universal Causal Deep Learning Models: The Case of Infinite-Dimensional Dynamical Systems from Stochastic Analysis , 2022, ArXiv.
[5] Ariel Neufeld,et al. Chaotic Hedging with Iterated Integrals and Neural Networks , 2022, ArXiv.
[6] Christa Cuchiero,et al. Universal approximation theorems for continuous functions of c\`adl\`ag paths and L\'evy-type signature models , 2022, 2208.02293.
[7] Yue Wang,et al. Neural Operator with Regularity Structure for Modeling Dynamics Driven by SPDEs , 2022, ArXiv.
[8] N. Matni,et al. Single Trajectory Nonparametric Learning of Nonlinear Dynamics , 2022, COLT.
[9] G Manjunath,et al. Embedding information onto a dynamical system , 2021, Nonlinearity.
[10] E. Weinan,et al. Representation formulas and pointwise properties for Barron functions , 2020, Calculus of Variations and Partial Differential Equations.
[11] E. Weinan,et al. Approximation and Optimization Theory for Linear Continuous-Time Recurrent Neural Networks , 2022, J. Mach. Learn. Res..
[12] B. Acciaio,et al. Metric Hypertransformers are Universal Adapted Maps , 2022, ArXiv.
[13] C. Salvi,et al. Neural Stochastic PDEs: Resolution-Invariant Learning of Continuous Spatiotemporal Dynamics , 2021, NeurIPS.
[14] Nikola B. Kovachki,et al. Neural Operator: Learning Maps Between Function Spaces , 2021, ArXiv.
[15] Juan-Pablo Ortega,et al. Learning strange attractors with reservoir systems , 2021, ArXiv.
[16] B. Hunt,et al. A Hybrid Approach to Atmospheric Modeling That Combines Machine Learning With a Physics‐Based Numerical Model , 2021, Journal of Advances in Modeling Earth Systems.
[17] Lukas Gonon,et al. Random feature neural networks learn Black-Scholes type PDEs without curse of dimensionality , 2021, ArXiv.
[18] Kohei Nakajima,et al. Learning Temporal Quantum Tomography , 2021, Physical review letters.
[19] E. Ott,et al. Using Data Assimilation to Train a Hybrid Forecast System that Combines Machine-Learning and Knowledge-Based Components , 2021, Chaos.
[20] Lukas Gonon,et al. Fading memory echo state networks are universal , 2020, Neural Networks.
[21] Nikola B. Kovachki,et al. Fourier Neural Operator for Parametric Partial Differential Equations , 2020, ICLR.
[22] Juan-Pablo Ortega,et al. Dimension reduction in recurrent networks by canonicalization , 2020, Journal of Geometric Mechanics.
[23] E Weinan,et al. Towards a Mathematical Understanding of Neural Network-Based Machine Learning: what we know and what we don't , 2020, CSIAM Transactions on Applied Mathematics.
[24] Kohei Nakajima,et al. Higher-Order Quantum Reservoir Computing , 2020, ArXiv.
[25] Anastasis Kratsios,et al. Non-Euclidean Universal Approximation , 2020, NeurIPS.
[26] E Weinan,et al. On the Banach spaces associated with multi-layer ReLU networks: Function representation, approximation theory and gradient descent dynamics , 2020, CSIAM Transactions on Applied Mathematics.
[27] Lukas Gonon,et al. Memory and forecasting capacities of nonlinear recurrent networks , 2020, Physica D: Nonlinear Phenomena.
[28] H. Nurdin,et al. Temporal Information Processing on Noisy Quantum Computers , 2020, Physical Review Applied.
[29] Peter Tino,et al. Dynamical Systems as Temporal Feature Spaces , 2019, J. Mach. Learn. Res..
[30] Juan-Pablo Ortega,et al. Reservoir Computing Universality With Stochastic Inputs , 2018, IEEE Transactions on Neural Networks and Learning Systems.
[31] G. Manjunath. Supplementary Material for “Stability and Memory-loss go hand-in-hand: Three Results in Dynamics & Computation” , 2020 .
[32] Wu Lei. A PRIORI ESTIMATES OF THE POPULATION RISK FOR TWO-LAYER NEURAL NETWORKS , 2020 .
[33] Lukas Gonon,et al. Risk bounds for reservoir computing , 2019, ArXiv.
[34] Juan-Pablo Ortega,et al. Differentiable reservoir computing , 2019, J. Mach. Learn. Res..
[35] Jiayin Chen,et al. Learning nonlinear input–output maps with dissipative quantum systems , 2019, Quantum Information Processing.
[36] Franz J. Király,et al. Kernels for sequentially ordered data , 2016, J. Mach. Learn. Res..
[37] Peter Tiño,et al. Asymptotic Fisher memory of randomized linear symmetric Echo State Networks , 2018, Neurocomputing.
[38] Juan-Pablo Ortega,et al. Echo state networks are universal , 2018, Neural Networks.
[39] Edward Ott,et al. Attractor reconstruction by machine learning. , 2018, Chaos.
[40] Jaideep Pathak,et al. Model-Free Prediction of Large Spatiotemporally Chaotic Systems from Data: A Reservoir Computing Approach. , 2018, Physical review letters.
[41] Juan-Pablo Ortega,et al. Universal discrete-time reservoir computers with stochastic inputs and linear readouts using non-homogeneous state-affine systems , 2017, J. Mach. Learn. Res..
[42] Jaideep Pathak,et al. Using machine learning to replicate chaotic attractors and calculate Lyapunov exponents from data. , 2017, Chaos.
[43] JAKE BOUVRIE,et al. Kernel Methods for the Approximation of Nonlinear Systems , 2011, SIAM J. Control. Optim..
[44] Hyunjoong Kim,et al. Functional Analysis I , 2017 .
[45] Romain Couillet,et al. The Asymptotic Performance of Linear Echo State Neural Networks , 2016, J. Mach. Learn. Res..
[46] Gábor Lugosi,et al. Concentration Inequalities - A Nonasymptotic Theory of Independence , 2013, Concentration Inequalities.
[47] Herbert Jaeger,et al. Echo State Property Linked to an Input: Exploring a Fundamental Characteristic of Recurrent Neural Networks , 2013, Neural Computation.
[48] Stefan J. Kiebel,et al. Re-visiting the echo state property , 2012, Neural Networks.
[49] Jake V. Bouvrie,et al. Kernel Methods for the Approximation of Some Key Quantities of Nonlinear Systems , 2012, 1204.0563.
[50] Benjamin Schrauwen,et al. Recurrent Kernel Machines: Computing with Infinite Echo State Networks , 2012, Neural Computation.
[51] Benjamin Schrauwen,et al. Memory in linear recurrent neural networks in continuous time , 2010, Neural Networks.
[52] S. Barry Cooper,et al. Computability In Context: Computation and Logic in the Real World , 2009 .
[53] C. Villani. Optimal Transport: Old and New , 2008 .
[54] Benjamin Recht,et al. Random Features for Large-Scale Kernel Machines , 2007, NIPS.
[55] Chee Kheong Siew,et al. Extreme learning machine: Theory and applications , 2006, Neurocomputing.
[56] Harald Haas,et al. Harnessing Nonlinearity: Predicting Chaotic Systems and Saving Energy in Wireless Communication , 2004, Science.
[57] Henry Markram,et al. Real-Time Computing Without Stable States: A New Framework for Neural Computation Based on Perturbations , 2002, Neural Computation.
[58] Herbert Jaeger,et al. The''echo state''approach to analysing and training recurrent neural networks , 2001 .
[59] Maxwell B. Stinchcombe,et al. Neural network approximation of continuous functionals and continuous functions on compactifications , 1999, Neural Networks.
[60] Hong Chen,et al. Approximation capability to functions of several variables, nonlinear functionals, and operators by radial basis function neural networks , 1993, IEEE Trans. Neural Networks.
[61] Andrew R. Barron,et al. Universal approximation bounds for superpositions of a sigmoidal function , 1993, IEEE Trans. Inf. Theory.
[62] Michael B. Matthews,et al. Approximating nonlinear fading-memory operators using neural network models , 1993 .
[63] Michael B. Matthews,et al. On the uniform approximation of nonlinear discrete-time fading-memory systems using neural network models , 1992 .
[64] Leon O. Chua,et al. Fading memory and the problem of approximating nonlinear operators with volterra series , 1985 .
[65] G. M.,et al. Partial Differential Equations I , 2023, Applied Mathematical Sciences.