Data-driven prediction of a multi-scale Lorenz 96 chaotic system using a hierarchy of deep learning methods: Reservoir computing, ANN, and RNN-LSTM

In this paper, the performance of three deep learning methods for predicting short-term evolution and reproducing the long-term statistics of a multi-scale spatio-temporal Lorenz 96 system is examined. The methods are: echo state network (a type of reservoir computing, RC-ESN), deep feed-forward artificial neural network (ANN), and recurrent neural network with long short-term memory (RNN-LSTM). This Lorenz system has three tiers of nonlinearly interacting variables representing slow/large-scale ($X$), intermediate ($Y$), and fast/small-scale ($Z$) processes. For training or testing, only $X$ is available; $Y$ and $Z$ are never known/used. It is shown that RC-ESN substantially outperforms ANN and RNN-LSTM for short-term prediction, e.g., accurately forecasting the chaotic trajectories for hundreds of numerical solver's time steps, equivalent to several Lyapunov timescales. RNN-LSTM and ANN show some prediction skills as well; RNN-LSTM bests ANN. Furthermore, even after losing the trajectory, data predicted by RC-ESN and RNN-LSTM have probability density functions (PDFs) that closely match the true PDF, even at the tails. PDF of the ANN data deviates from the true PDF. Implications, caveats, and applications to data-driven and inexact, data-assisted surrogate modeling of complex dynamical systems such as weather/climate are discussed.

[1]  Pierre Gentine,et al.  Could Machine Learning Break the Convection Parameterization Deadlock? , 2018, Geophysical Research Letters.

[2]  Peter D. Düben,et al.  On the use of scale‐dependent precision in Earth System modelling , 2017 .

[3]  C. Jones,et al.  Development and evaluation of an Earth-System model - HadGEM2 , 2011 .

[4]  Andrew Gettelman,et al.  The Art and Science of Climate Model Tuning , 2017 .

[5]  Petros Koumoutsakos,et al.  Data-driven forecasting of high-dimensional chaotic systems with long short-term memory networks , 2018, Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences.

[6]  Rustam M. Vahidov,et al.  Application of machine learning techniques for supply chain demand forecasting , 2008, Eur. J. Oper. Res..

[7]  Jaideep Pathak,et al.  Using machine learning to replicate chaotic attractors and calculate Lyapunov exponents from data. , 2017, Chaos.

[8]  Rolando R. Garcia,et al.  Modification of the Gravity Wave Parameterization in the Whole Atmosphere Community Climate Model: Motivation and Results , 2017 .

[9]  Sebastian Scher,et al.  Generalization properties of feed-forward neural networks trained on Lorenz systems , 2019, Nonlinear Processes in Geophysics.

[10]  Igor Mezic,et al.  Ergodic Theory, Dynamic Mode Decomposition, and Computation of Spectral Properties of the Koopman Operator , 2016, SIAM J. Appl. Dyn. Syst..

[11]  Jürgen Schmidhuber,et al.  Long Short-Term Memory , 1997, Neural Computation.

[12]  Steven L. Brunton,et al.  On dynamic mode decomposition: Theory and applications , 2013, 1312.0041.

[13]  J. Nathan Kutz,et al.  Deep learning in fluid dynamics , 2017, Journal of Fluid Mechanics.

[14]  Hal Finkel,et al.  Doing Moore with Less - Leapfrogging Moore's Law with Inexactness for Supercomputing , 2016, ArXiv.

[15]  Saleh Nabi,et al.  Reduced-order modeling of fully turbulent buoyancy-driven flows using the Green's function method , 2018, Physical Review Fluids.

[16]  W. Marsden I and J , 2012 .

[17]  D. Randall,et al.  A cloud resolving model as a cloud parameterization in the NCAR Community Climate System Model: Preliminary results , 2001 .

[18]  Razvan Pascanu,et al.  On the difficulty of training recurrent neural networks , 2012, ICML.

[19]  W. R. Peltier,et al.  Deep learning of mixing by two ‘atoms’ of stratified turbulence , 2018, Journal of Fluid Mechanics.

[20]  Zhizhen Zhao,et al.  Spatiotemporal Pattern Extraction by Spectral Analysis of Vector-Valued Observables , 2017, Journal of Nonlinear Science.

[21]  Garrison W. Cottrell,et al.  WALKING WALKing walking: Action Recognition from Action Echoes , 2017, IJCAI.

[22]  Quoc V. Le,et al.  Sequence to Sequence Learning with Neural Networks , 2014, NIPS.

[23]  Guigang Zhang,et al.  Deep Learning , 2016, Int. J. Semantic Comput..

[24]  Hava T. Siegelmann,et al.  On the computational power of neural nets , 1992, COLT '92.

[25]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[26]  Geoffrey E. Hinton,et al.  Speech recognition with deep recurrent neural networks , 2013, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.

[27]  A. Mohan,et al.  Compressed Convolutional LSTM: An Efficient Deep Learning framework to Model High Fidelity 3D Turbulence , 2019, 1903.00033.

[28]  W. Collins,et al.  The Formulation and Atmospheric Simulation of the Community Atmosphere Model Version 3 (CAM3) , 2006 .

[29]  Petros Koumoutsakos,et al.  Data-assisted reduced-order modeling of extreme events in complex dynamical systems , 2018, PloS one.

[30]  Christopher K. Wikle,et al.  An ensemble quadratic echo state network for non‐linear spatio‐temporal forecasting , 2017, 1708.05094.

[31]  N. MacDonald Nonlinear dynamics , 1980, Nature.

[32]  Karthik Kashinath,et al.  Deep Learning for Scientific Inference from Geophysical Data: The Madden-Julian Oscillation as a Test Case , 2019, 1902.04621.

[33]  C. Caramanis What is ergodic theory , 1963 .

[34]  S. Bony,et al.  What Are Climate Models Missing? , 2013, Science.

[35]  Clarence W. Rowley,et al.  A Data–Driven Approximation of the Koopman Operator: Extending Dynamic Mode Decomposition , 2014, Journal of Nonlinear Science.

[36]  I. Mezić Spectral Properties of Dynamical Systems, Model Reduction and Decompositions , 2005 .

[37]  Peter D. Düben,et al.  Improving Weather Forecast Skill through Reduced-Precision Data Assimilation , 2018 .

[38]  G. Flato Earth system models: an overview , 2011 .

[39]  Karthik Kashinath,et al.  Enforcing statistical constraints in generative adversarial networks for modeling chaotic dynamical systems , 2019, J. Comput. Phys..

[40]  Edward Ott,et al.  Attractor reconstruction by machine learning. , 2018, Chaos.

[41]  Kai Chen,et al.  A LSTM-based method for stock returns prediction: A case study of China stock market , 2015, 2015 IEEE International Conference on Big Data (Big Data).

[42]  Hannah M. Christensen,et al.  Machine Learning for Stochastic Parameterization: Generative Adversarial Networks in the Lorenz '96 Model , 2019, Journal of Advances in Modeling Earth Systems.

[43]  M. Pritchard,et al.  Rainfall From Resolved Rather Than Parameterized Processes Better Represents the Present‐Day and Climate Change Response of Moderate Rates in the Community Atmosphere Model , 2018, Journal of advances in modeling earth systems.

[44]  Matthew Chantry,et al.  Scale-Selective Precision for Weather and Climate Forecasting , 2019, Monthly Weather Review.

[45]  Tim Palmer,et al.  Climate forecasting: Build high-resolution global climate models , 2014, Nature.

[46]  Dit-Yan Yeung,et al.  Convolutional LSTM Network: A Machine Learning Approach for Precipitation Nowcasting , 2015, NIPS.

[47]  J. Salas,et al.  Nonlinear dynamics, delay times, and embedding windows , 1999 .

[48]  Bozhkov Lachezar,et al.  Echo State Network , 2017, Encyclopedia of Machine Learning and Data Mining.

[49]  Paris Perdikaris,et al.  Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations , 2019, J. Comput. Phys..

[50]  A. P. Siebesma,et al.  Climate goals and computing the future of clouds , 2017 .

[51]  Noah D. Brenowitz,et al.  Prognostic Validation of a Neural Network Unified Physics Parameterization , 2018, Geophysical Research Letters.

[52]  Woosok Moon,et al.  Predicting Rare Events in Multiscale Dynamical Systems using Machine Learning , 2019, 1908.03771.

[53]  John Augustine,et al.  Opportunities for energy efficient computing: A study of inexact general purpose processors for high-performance and big-data applications , 2015, 2015 Design, Automation & Test in Europe Conference & Exhibition (DATE).

[54]  Adrian Sandu,et al.  A Machine Learning Approach to Adaptive Covariance Localization , 2018, ArXiv.

[55]  Steven L. Brunton,et al.  Data-Driven Science and Engineering , 2019 .

[56]  Andrew Stuart,et al.  Earth System Modeling 2.0: A Blueprint for Models That Learn From Observations and Targeted High‐Resolution Simulations , 2017, 1709.00037.

[57]  Joachim Denzler,et al.  Deep learning and process understanding for data-driven Earth system science , 2019, Nature.

[58]  E. Lorenz Predictability of Weather and Climate: Predictability – a problem partly solved , 2006 .

[59]  R. Brockett,et al.  Reservoir observers: Model-free inference of unmeasured variables in chaotic systems. , 2017, Chaos.

[60]  K. Palem,et al.  Inexactness and a future of computing , 2014, Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences.

[61]  Peter D. Düben,et al.  Benchmark Tests for Numerical Weather Forecasts on Inexact Hardware , 2014 .

[62]  V. A. Epanechnikov Non-Parametric Estimation of a Multivariate Probability Density , 1969 .

[63]  Peter A. G. Watson,et al.  Applying Machine Learning to Improve Simulations of a Chaotic Dynamical System Using Empirical Error Correction , 2019, Journal of advances in modeling earth systems.

[64]  M. Vastaranta,et al.  This is a non-peer reviewed preprint submitted to EarthArXiv , 2020 .

[65]  Ulrich Parlitz,et al.  Observing spatio-temporal dynamics of excitable media using reservoir computing. , 2018, Chaos.

[66]  Herbert Jaeger,et al.  Reservoir computing approaches to recurrent neural network training , 2009, Comput. Sci. Rev..

[67]  Thomas Bolton,et al.  Applications of Deep Learning to Ocean Data Inference and Subgrid Parameterization , 2019, Journal of Advances in Modeling Earth Systems.

[68]  Z. Kuang,et al.  Moist Static Energy Budget of MJO-like Disturbances in the Atmosphere of a Zonally Symmetric Aquaplanet , 2012 .

[69]  H. S. Kim,et al.  Nonlinear dynamics , delay times , and embedding windows , 1999 .

[70]  Yisong Yue,et al.  Long-term Forecasting using Tensor-Train RNNs , 2017, ArXiv.

[71]  Peter Bauer,et al.  The quiet revolution of numerical weather prediction , 2015, Nature.

[72]  Peter D. Düben,et al.  On the use of inexact, pruned hardware in atmospheric modelling , 2014, Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences.

[73]  Paris Perdikaris,et al.  Physics-Constrained Deep Learning for High-dimensional Surrogate Modeling and Uncertainty Quantification without Labeled Data , 2019, J. Comput. Phys..

[74]  Olivier Temam,et al.  Leveraging the error resilience of machine-learning applications for designing highly energy efficient accelerators , 2014, 2014 19th Asia and South Pacific Design Automation Conference (ASP-DAC).

[75]  P. O'Gorman,et al.  Using Machine Learning to Parameterize Moist Convection: Potential for Modeling of Climate, Climate Change, and Extreme Events , 2018, Journal of Advances in Modeling Earth Systems.

[76]  Pedram Hassanzadeh,et al.  Data-driven reduced modelling of turbulent Rayleigh–Bénard convection using DMD-enhanced fluctuation–dissipation theorem , 2018, Journal of Fluid Mechanics.

[77]  Steven L. Brunton,et al.  Deep learning of dynamics and signal-noise decomposition with time-stepping constraints , 2018, J. Comput. Phys..

[78]  Peter Bauer,et al.  Challenges and design choices for global weather and climate models based on machine learning , 2018, Geoscientific Model Development.

[79]  Karthik Duraisamy,et al.  Turbulence Modeling in the Age of Data , 2018, Annual Review of Fluid Mechanics.

[80]  P. Hassanzadeh,et al.  A perspective on climate model hierarchies , 2017 .

[81]  Radford M. Neal Pattern Recognition and Machine Learning , 2007, Technometrics.

[82]  Tie-Yan Liu,et al.  Convergence Analysis of Distributed Stochastic Gradient Descent with Shuffling , 2017, Neurocomputing.

[83]  Michelle Girvan,et al.  Hybrid Forecasting of Chaotic Processes: Using Machine Learning in Conjunction with a Knowledge-Based Model , 2018, Chaos.

[84]  Stefan J. Kiebel,et al.  Re-visiting the echo state property , 2012, Neural Networks.

[85]  Jaideep Pathak,et al.  Model-Free Prediction of Large Spatiotemporally Chaotic Systems from Data: A Reservoir Computing Approach. , 2018, Physical review letters.

[86]  David A. Randall,et al.  Structure of the Madden-Julian Oscillation in the Superparameterized CAM , 2009 .

[87]  J. Templeton,et al.  Reynolds averaged turbulence modelling using deep neural networks with embedded invariance , 2016, Journal of Fluid Mechanics.

[88]  Yoshua Bengio,et al.  Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.

[89]  Harald Haas,et al.  Harnessing Nonlinearity: Predicting Chaotic Systems and Saving Energy in Wireless Communication , 2004, Science.

[90]  Pierre Gentine,et al.  Deep learning to represent subgrid processes in climate models , 2018, Proceedings of the National Academy of Sciences.