On Theory-training Neural Networks to Infer the Solution of Highly Coupled Differential Equations

Deep neural networks are transforming fields ranging from computer vision to computational medicine, and we recently extended their application to the field of phase-change heat transfer by introducing theory-trained neural networks (TTNs) for a solidification problem [1]. Here, we present general, in-depth, and empirical insights into theory-training networks for learning the solution of highly coupled differential equations. We analyze the deteriorating effects of the oscillating loss on the ability of a network to satisfy the equations at the training data points, measured by the final training loss, and on the accuracy of the inferred solution. We introduce a theory-training technique that, by leveraging regularization, eliminates those oscillations, decreases the final training loss, and improves the accuracy of the inferred solution, with no additional computational cost. Then, we present guidelines that allow a systematic search for the network that has the optimal training time and inference accuracy for a given set of equations; following these guidelines can reduce the number of tedious training iterations in that search. Finally, a comparison between theory-training and the rival, conventional method of solving differential equations using discretization attests to the advantages of theory-training not being necessarily limited to high-dimensional sets of equations. The comparison also reveals a limitation of the current theory-training framework that may limit its application in domains where extreme accuracies are necessary.

[1]  Paris Perdikaris,et al.  Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations , 2019, J. Comput. Phys..

[2]  Ruben Juanes,et al.  A deep learning framework for solution and discovery in solid mechanics , 2020 .

[3]  Maziar Raissi,et al.  Deep Hidden Physics Models: Deep Learning of Nonlinear Partial Differential Equations , 2018, J. Mach. Learn. Res..

[4]  Maziar Raissi,et al.  Forward-Backward Stochastic Neural Networks: Deep Learning of High-dimensional Partial Differential Equations , 2018, ArXiv.

[5]  George E. Karniadakis,et al.  Hidden physics models: Machine learning of nonlinear partial differential equations , 2017, J. Comput. Phys..

[6]  Boaz Barak,et al.  Deep double descent: where bigger models and more data hurt , 2019, ICLR.

[7]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[8]  R. Juanes,et al.  SciANN: A Keras/TensorFlow wrapper for scientific computations and physics-informed deep learning using artificial neural networks , 2020, Computer Methods in Applied Mechanics and Engineering.

[9]  C. Beckermann,et al.  A truncated-Scheil-type model for columnar solidification of binary alloys in the presence of melt convection , 2019, Materialia.

[10]  C. Beckermann,et al.  Validation of a Model for the Columnar to Equiaxed Transition with Melt Convection , 2016 .

[11]  G. J. Schmitz,et al.  Theory-training deep neural networks for an alloy solidification benchmark problem , 2019, Computational Materials Science.

[12]  George Em Karniadakis,et al.  Quantifying the generalization error in deep learning in terms of data distribution and neural network smoothness , 2019, Neural Networks.

[13]  George Em Karniadakis,et al.  Locally adaptive activation functions with slope recovery for deep and physics-informed neural networks , 2020, Proceedings of the Royal Society A.

[14]  Michael S. Triantafyllou,et al.  Deep learning of vortex-induced vibrations , 2018, Journal of Fluid Mechanics.

[15]  George Em Karniadakis,et al.  Adaptive activation functions accelerate convergence in deep and physics-informed neural networks , 2019, J. Comput. Phys..

[16]  C. Beckermann,et al.  Rayleigh Number Criterion for Formation of A-Segregates in Steel Castings and Ingots , 2013, Metallurgical and Materials Transactions A.

[17]  Yibo Yang,et al.  Physics-Informed Neural Networks for Cardiac Activation Mapping , 2020, Frontiers in Physics.

[18]  L. R. Scott,et al.  The Mathematical Theory of Finite Element Methods , 1994 .