Quantifying total uncertainty in physics-informed neural networks for solving forward and inverse stochastic problems

Abstract Physics-informed neural networks (PINNs) have recently emerged as an alternative way of numerically solving partial differential equations (PDEs) without the need of building elaborate grids, instead, using a straightforward implementation. In particular, in addition to the deep neural network (DNN) for the solution, an auxiliary DNN is considered that represents the residual of the PDE. The residual is then combined with the mismatch in the given data of the solution in order to formulate the loss function. This framework is effective but is lacking uncertainty quantification of the solution due to the inherent randomness in the data or due to the approximation limitations of the DNN architecture. Here, we propose a new method with the objective of endowing the DNN with uncertainty quantification for both sources of uncertainty, i.e., the parametric uncertainty and the approximation uncertainty. We first account for the parametric uncertainty when the parameter in the differential equation is represented as a stochastic process. Multiple DNNs are designed to learn the modal functions of the arbitrary polynomial chaos (aPC) expansion of its solution by using stochastic data from sparse sensors. We can then make predictions from new sensor measurements very efficiently with the trained DNNs. Moreover, we employ dropout to quantify the uncertainty of DNNs in approximating the modal functions. We then design an active learning strategy based on the dropout uncertainty to place new sensors in the domain in order to improve the predictions of DNNs. Several numerical tests are conducted for both the forward and the inverse problems to demonstrate the effectiveness of PINNs combined with uncertainty quantification. This NN-aPC new paradigm of physics-informed deep learning with uncertainty quantification can be readily applied to other types of stochastic PDEs in multi-dimensions.

[1]  Michael I. Jordan,et al.  Variational Bayesian Inference with Stochastic Search , 2012, ICML.

[2]  G. Karniadakis,et al.  Multi-Element Generalized Polynomial Chaos for Arbitrary Probability Measures , 2006, SIAM J. Sci. Comput..

[3]  Max Welling,et al.  Auto-Encoding Variational Bayes , 2013, ICLR.

[4]  Andrew M. Stuart,et al.  Inverse problems: A Bayesian perspective , 2010, Acta Numerica.

[5]  Roberto Cipolla,et al.  Bayesian SegNet: Model Uncertainty in Deep Convolutional Encoder-Decoder Architectures for Scene Understanding , 2015, BMVC.

[6]  Sergey Oladyshkin,et al.  Data-driven uncertainty quantification using the arbitrary polynomial chaos expansion , 2012, Reliab. Eng. Syst. Saf..

[7]  Léon Personnaz,et al.  Construction of confidence intervals for neural networks based on least squares estimation , 2000, Neural Networks.

[8]  Geoffrey E. Hinton,et al.  Bayesian Learning for Neural Networks , 1995 .

[9]  Liu Yang,et al.  Neural-net-induced Gaussian process regression for function approximation and PDE solution , 2018, J. Comput. Phys..

[10]  Nitish Srivastava,et al.  Improving neural networks by preventing co-adaptation of feature detectors , 2012, ArXiv.

[11]  Julien Cornebise,et al.  Weight Uncertainty in Neural Networks , 2015, ArXiv.

[12]  Hadi Meidani,et al.  A Deep Neural Network Surrogate for High-Dimensional Random Partial Differential Equations , 2018, Probabilistic Engineering Mechanics.

[13]  Suzanne Hurter,et al.  Heat flow from the Earth's interior: Analysis of the global data set , 1993 .

[14]  Yuanbo Liu,et al.  Soil Salinity Retrieval from Advanced Multi-Spectral Sensor with Partial Least Square Regression , 2015, Remote. Sens..

[15]  David J. C. MacKay,et al.  A Practical Bayesian Framework for Backpropagation Networks , 1992, Neural Computation.

[16]  Lexing Ying,et al.  Solving parametric PDE problems with artificial neural networks , 2017, European Journal of Applied Mathematics.

[17]  Xiu Yang,et al.  Physics-Informed Kriging: A Physics-Informed Gaussian Process Regression Method for Data-Model Convergence , 2018, ArXiv.

[18]  Michael I. Jordan,et al.  An Introduction to Variational Methods for Graphical Models , 1999, Machine-mediated learning.

[19]  Paris Perdikaris,et al.  Learning Parameters and Constitutive Relationships with Physics Informed Deep Neural Networks , 2018, 1808.03398.

[20]  Chong Wang,et al.  Stochastic variational inference , 2012, J. Mach. Learn. Res..

[21]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[22]  Nitish Srivastava,et al.  Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..

[23]  J. York,et al.  Bayesian Graphical Models for Discrete Data , 1995 .

[24]  George Em Karniadakis,et al.  Adaptive multi-element polynomial chaos with discrete measure , 2015 .

[25]  Dongbin Xiu,et al.  The Wiener-Askey Polynomial Chaos for Stochastic Differential Equations , 2002, SIAM J. Sci. Comput..

[26]  Steven L. Brunton,et al.  Data-Driven Identification of Parametric Partial Differential Equations , 2018, SIAM J. Appl. Dyn. Syst..

[27]  Maziar Raissi,et al.  Forward-Backward Stochastic Neural Networks: Deep Learning of High-dimensional Partial Differential Equations , 2018, ArXiv.

[28]  Thore Graepel,et al.  Solving Noisy Linear Operator Equations by Gaussian Processes: Application to Ordinary and Partial Differential Equations , 2003, ICML.

[29]  Thomas Y. Hou,et al.  A dynamically bi-orthogonal method for time-dependent stochastic partial differential equations II: Adaptivity and generalizations , 2013, J. Comput. Phys..

[30]  Mark E. Borsuk,et al.  Improving Structure MCMC for Bayesian Networks through Markov Blanket Resampling , 2016, J. Mach. Learn. Res..

[31]  Paris Perdikaris,et al.  Physics Informed Deep Learning (Part II): Data-driven Discovery of Nonlinear Partial Differential Equations , 2017, ArXiv.

[32]  Daan Wierstra,et al.  Stochastic Backpropagation and Approximate Inference in Deep Generative Models , 2014, ICML.

[33]  Pierre F. J. Lermusiaux,et al.  Dynamically orthogonal field equations for continuous stochastic dynamical systems , 2009 .

[34]  Nir Friedman,et al.  Probabilistic Graphical Models - Principles and Techniques , 2009 .

[35]  Steven L. Brunton,et al.  Data-driven discovery of partial differential equations , 2016, Science Advances.

[36]  Zoubin Ghahramani,et al.  A Theoretically Grounded Application of Dropout in Recurrent Neural Networks , 2015, NIPS.

[37]  Alex Kendall,et al.  Concrete Dropout , 2017, NIPS.

[38]  R. Ghanem,et al.  Polynomial Chaos in Stochastic Finite Elements , 1990 .

[39]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.

[40]  Xiao Yang,et al.  Fast Predictive Image Registration , 2016, LABELS/DLMIA@MICCAI.

[41]  Miguel Lázaro-Gredilla,et al.  Doubly Stochastic Variational Bayes for non-Conjugate Inference , 2014, ICML.

[42]  Jing Li,et al.  A data-driven framework for sparsity-enhanced surrogates with arbitrary mutually dependent randomness. , 2018, Computer methods in applied mechanics and engineering.

[43]  George E. Karniadakis,et al.  Hidden physics models: Machine learning of nonlinear partial differential equations , 2017, J. Comput. Phys..

[44]  Zoubin Ghahramani,et al.  Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning , 2015, ICML.

[45]  Yuan Yu,et al.  TensorFlow: A system for large-scale machine learning , 2016, OSDI.

[46]  E Weinan,et al.  Deep Learning-Based Numerical Methods for High-Dimensional Parabolic Partial Differential Equations and Backward Stochastic Differential Equations , 2017, Communications in Mathematics and Statistics.

[47]  Jeroen A. S. Witteveen,et al.  Modeling Arbitrary Uncertainties Using Gram-Schmidt Polynomial Chaos , 2006 .

[48]  Neil D. Lawrence,et al.  Deep Gaussian Processes , 2012, AISTATS.

[49]  Zoubin Ghahramani,et al.  Variational Bayesian dropout: pitfalls and fixes , 2018, ICML.

[50]  Ilias Bilionis,et al.  Probabilistic solvers for partial differential equations , 2016, 1607.03526.

[51]  Hadi Meidani,et al.  A deep learning solution approach for high-dimensional random differential equations , 2019, Probabilistic Engineering Mechanics.

[52]  Thomas Y. Hou,et al.  A dynamically bi-orthogonal method for time-dependent stochastic partial differential equations I: Derivation and algorithms , 2013, J. Comput. Phys..

[53]  Nicholas Zabaras,et al.  Bayesian Deep Convolutional Encoder-Decoder Networks for Surrogate Modeling and Uncertainty Quantification , 2018, J. Comput. Phys..

[54]  Yarin Gal,et al.  Dropout Inference in Bayesian Neural Networks with Alpha-divergences , 2017, ICML.

[55]  Liu Yang,et al.  Physics-Informed Generative Adversarial Networks for Stochastic Differential Equations , 2018, SIAM J. Sci. Comput..

[56]  Paris Perdikaris,et al.  Numerical Gaussian Processes for Time-Dependent and Nonlinear Partial Differential Equations , 2017, SIAM J. Sci. Comput..

[57]  Dimitrios I. Fotiadis,et al.  Artificial neural networks for solving ordinary and partial differential equations , 1997, IEEE Trans. Neural Networks.

[58]  Paris Perdikaris,et al.  Physics Informed Deep Learning (Part I): Data-driven Solutions of Nonlinear Partial Differential Equations , 2017, ArXiv.

[59]  Paris Perdikaris,et al.  Machine learning of linear differential equations using Gaussian processes , 2017, J. Comput. Phys..

[60]  Simo Särkkä,et al.  Linear Operators and Stochastic Partial Differential Equations in Gaussian Process Regression , 2011, ICANN.

[61]  O. Stegle,et al.  DeepCpG: accurate prediction of single-cell DNA methylation states using deep learning , 2016, Genome Biology.

[62]  Alex Kendall,et al.  What Uncertainties Do We Need in Bayesian Deep Learning for Computer Vision? , 2017, NIPS.

[63]  Dimitris G. Papageorgiou,et al.  Neural-network methods for boundary value problems with irregular boundaries , 2000, IEEE Trans. Neural Networks Learn. Syst..