Deep neural network for solving differential equations motivated by Legendre-Galerkin approximation

Nonlinear differential equations are challenging to solve numerically and are important to understanding the dynamics of many physical systems. Deep neural networks have been applied to help alleviate the computational cost that is associated with solving these systems. We explore the performance and accuracy of various neural architectures on both linear and nonlinear differential equations by creating accurate training sets with the spectral element method. Next, we implement a novel Legendre-Galerkin Deep Neural Network (LGNet) algorithm to predict solutions to differential equations. By constructing a set of a linear combination of the Legendre basis, we predict the corresponding coefficients, $\alpha_i$ which successfully approximate the solution as a sum of smooth basis functions $u \simeq \sum_{i=0}^{N} \alpha_i \varphi_i$. As a computational example, linear and nonlinear models with Dirichlet or Neumann boundary conditions are considered.

[1]  Yoon Kim,et al.  Convolutional Neural Networks for Sentence Classification , 2014, EMNLP.

[2]  Steve B. Jiang,et al.  Nonlinear Systems Identification Using Deep Dynamic Neural Networks , 2016, ArXiv.

[3]  Roger Temam,et al.  Enriched numerical scheme for singularly perturbed barotropic Quasi-Geostrophic equations , 2020, J. Comput. Phys..

[4]  Stephan Hoyer,et al.  Learned discretizations for passive scalar advection in a 2-D turbulent flow , 2020 .

[5]  R. Tibshirani Regression Shrinkage and Selection via the Lasso , 1996 .

[6]  Jean-François Richard,et al.  Methods of Numerical Integration , 2000 .

[7]  Kurt Hornik,et al.  Multilayer feedforward networks are universal approximators , 1989, Neural Networks.

[8]  Satish C. Reddy,et al.  The Accuracy of the Chebyshev Differencing Method for Analytic Functions , 2005, SIAM J. Numer. Anal..

[9]  Silvia Ferrari,et al.  A constrained integration (CINT) approach to solving partial differential equations using artificial neural networks , 2015, Neurocomputing.

[10]  Sergey Ioffe,et al.  Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.

[11]  Roger Temam,et al.  On the numerical approximations of stiff convection–diffusion equations in a circle , 2014, Numerische Mathematik.

[12]  Youngjoon Hong,et al.  A high-order perturbation of surfaces method for vector electromagnetic scattering by doubly layered periodic crossed gratings , 2018, J. Comput. Phys..

[13]  Raman Arora,et al.  Understanding Deep Neural Networks with Rectified Linear Units , 2016, Electron. Colloquium Comput. Complex..

[14]  Vladimir N. Vapnik,et al.  The Nature of Statistical Learning Theory , 2000, Statistics for Engineering and Information Science.

[15]  Quoc V. Le,et al.  Searching for Activation Functions , 2018, arXiv.

[16]  Sebastian Ruder,et al.  An overview of gradient descent optimization algorithms , 2016, Vestnik komp'iuternykh i informatsionnykh tekhnologii.

[17]  George Em Karniadakis,et al.  Hidden fluid mechanics: Learning velocity and pressure fields from flow visualizations , 2020, Science.

[18]  George Cybenko,et al.  Approximation by superpositions of a sigmoidal function , 1989, Math. Control. Signals Syst..

[19]  George E. Karniadakis,et al.  Hidden physics models: Machine learning of nonlinear partial differential equations , 2017, J. Comput. Phys..

[20]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[21]  Yoshua Bengio,et al.  Gradient-based learning applied to document recognition , 1998, Proc. IEEE.

[22]  E. G. Steward,et al.  Fourier Optics: An Introduction, Second Edition. , 1989 .

[23]  Youngjoon Hong,et al.  A high-order perturbation of surfaces method for scattering of linear waves by periodic multiply layered gratings in two and three dimensions , 2017, J. Comput. Phys..

[24]  G. E. Karniadakis,et al.  Variational Physics-Informed Neural Networks For Solving Partial Differential Equations , 2019, ArXiv.

[25]  H. White,et al.  Universal approximation using feedforward networks with non-sigmoid hidden layer activation functions , 1989, International 1989 Joint Conference on Neural Networks.

[26]  Lu Liu,et al.  Rethinking 1D-CNN for Time Series Classification: A Stronger Baseline , 2020, ArXiv.

[27]  Radford M. Neal Pattern Recognition and Machine Learning , 2007, Technometrics.

[28]  Youngjoon Hong,et al.  Robust Neural Networks inspired by Strong Stability Preserving Runge-Kutta methods , 2020, ECCV.

[29]  Paris Perdikaris,et al.  Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations , 2019, J. Comput. Phys..

[30]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[31]  Jianfeng Gao,et al.  Scalable training of L1-regularized log-linear models , 2007, ICML '07.

[32]  Yixin Chen,et al.  Multi-Scale Convolutional Neural Networks for Time Series Classification , 2016, ArXiv.

[33]  David M. Allen,et al.  The Relationship Between Variable Selection and Data Agumentation and a Method for Prediction , 1974 .

[34]  Heike Freud,et al.  On Line Learning In Neural Networks , 2016 .

[35]  D. Gottlieb,et al.  Numerical analysis of spectral methods , 1977 .

[36]  Guigang Zhang,et al.  Deep Learning , 2016, Int. J. Semantic Comput..

[37]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.