Multi-fidelity modeling using DGPs: Improvements and a generalization to varying input space dimensions

Multi-fidelity approaches improve the inference of a high-fidelity model which is constructed using a small set of accurate observations, by taking advantage of its correlations with a low-fidelity model built using a larger set of approximated data. Most existing multi-fidelity methods consider the inputs of the low and high fidelity models defined identically over the same input space. However, it happens that the low fidelity model variables are defined over a different space than the variables of the high fidelity model due to different modeling approaches i.e. input spaces with different dimensionality and different nature of the variables. Recently, Deep Gaussian Processes have been used to exhibit the correlations between the low and high fidelity models. In this paper, Deep Gaussian Processes for multi-fidelity (MF-DGP) are extended to the case where the input spaces of the low and high fidelity models are different. Moreover, the learning capacity of MF-DGP is improved by proposing an optimization approach for the inducing inputs and by using natural gradients for the variational distributions of the inducing variables which also allows time reduction in the training.

[1]  Neil D. Lawrence,et al.  Deep Gaussian Processes for Multi-fidelity Modeling , 2019, ArXiv.

[2]  Marc Peter Deisenroth,et al.  Doubly Stochastic Variational Inference for Deep Gaussian Processes , 2017, NIPS.

[3]  Shifeng Xiong,et al.  Sequential Design and Analysis of High-Accuracy and Low-Accuracy Computer Codes , 2013, Technometrics.

[4]  Neil D. Lawrence,et al.  Gaussian Processes for Big Data , 2013, UAI.

[5]  Neil D. Lawrence,et al.  Deep Gaussian Processes , 2012, AISTATS.

[6]  A. O'Hagan,et al.  Bayesian calibration of computer models , 2001 .

[7]  Andreas C. Damianou,et al.  Nonlinear information fusion algorithms for data-efficient multi-fidelity modelling , 2017, Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences.

[8]  Mark Pullin,et al.  Emulation of physical processes with Emukit , 2021, ArXiv.

[9]  James Hensman,et al.  Natural Gradients in Practice: Non-Conjugate Variational Inference in Gaussian Process Models , 2018, AISTATS.

[10]  Shun-ichi Amari,et al.  Why natural gradient? , 1998, Proceedings of the 1998 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP '98 (Cat. No.98CH36181).

[11]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[12]  Loic Le Gratiet,et al.  RECURSIVE CO-KRIGING MODEL FOR DESIGN OF COMPUTER EXPERIMENTS WITH MULTIPLE LEVELS OF FIDELITY , 2012, 1210.0686.

[13]  Wei Chen,et al.  Input Mapping for Model Calibration with Application to Wing Aerodynamics , 2019, AIAA Journal.