Gaussian Processes and Polynomial Chaos Expansion for Regression Problem: Linkage via the RKHS and Comparison via the KL Divergence

In this paper, we examine two widely-used approaches, the polynomial chaos expansion (PCE) and Gaussian process (GP) regression, for the development of surrogate models. The theoretical differences between the PCE and GP approximations are discussed. A state-of-the-art PCE approach is constructed based on high precision quadrature points; however, the need for truncation may result in potential precision loss; the GP approach performs well on small datasets and allows a fine and precise trade-off between fitting the data and smoothing, but its overall performance depends largely on the training dataset. The reproducing kernel Hilbert space (RKHS) and Mercer’s theorem are introduced to form a linkage between the two methods. The theorem has proven that the two surrogates can be embedded in two isomorphic RKHS, by which we propose a novel method named Gaussian process on polynomial chaos basis (GPCB) that incorporates the PCE and GP. A theoretical comparison is made between the PCE and GPCB with the help of the Kullback–Leibler divergence. We present that the GPCB is as stable and accurate as the PCE method. Furthermore, the GPCB is a one-step Bayesian method that chooses the best subset of RKHS in which the true function should lie, while the PCE method requires an adaptive procedure. Simulations of 1D and 2D benchmark functions show that GPCB outperforms both the PCE and classical GP methods. In order to solve high dimensional problems, a random sample scheme with a constructive design (i.e., tensor product of quadrature points) is proposed to generate a valid training dataset for the GPCB method. This approach utilizes the nature of the high numerical accuracy underlying the quadrature points while ensuring the computational feasibility. Finally, the experimental results show that our sample strategy has a higher accuracy than classical experimental designs; meanwhile, it is suitable for solving high dimensional problems.

[1]  Carl E. Rasmussen,et al.  A Unifying View of Sparse Approximate Gaussian Process Regression , 2005, J. Mach. Learn. Res..

[2]  N. Aronszajn Theory of Reproducing Kernels. , 1950 .

[3]  D. Xiu Numerical Methods for Stochastic Computations: A Spectral Method Approach , 2010 .

[4]  Dongbin Xiu,et al.  High-Order Collocation Methods for Differential Equations with Random Inputs , 2005, SIAM J. Sci. Comput..

[5]  Iain Murray Introduction To Gaussian Processes , 2008 .

[6]  D. Higdon,et al.  Computer Model Calibration Using High-Dimensional Output , 2008 .

[7]  Hans-Paul Schwefel,et al.  Evolution and Optimum Seeking: The Sixth Generation , 1993 .

[8]  R. Ghanem,et al.  Stochastic Finite Elements: A Spectral Approach , 1990 .

[9]  Jian Wang,et al.  LIF: A new Kriging based learning function and its application to structural reliability analysis , 2017, Reliab. Eng. Syst. Saf..

[10]  Osman Balci Verification, validation, and certification of modeling and simulation applications , 2003, Proceedings of the 2003 Winter Simulation Conference, 2003..

[11]  T. Ishigami,et al.  An importance quantification technique in uncertainty analysis for computer models , 1990, [1990] Proceedings. First International Symposium on Uncertainty Modeling and Analysis.

[12]  Alex H. Barbat,et al.  Monte Carlo techniques in computational stochastic mechanics , 1998 .

[13]  V. Dubourg Adaptive surrogate models for reliability analysis and reliability-based design optimization , 2011 .

[14]  A. O'Hagan,et al.  Bayesian emulation of complex multi-output and dynamic computer models , 2010 .

[15]  Carl E. Rasmussen,et al.  Gaussian processes for machine learning , 2005, Adaptive computation and machine learning.

[16]  J. Wiart,et al.  Polynomial-Chaos-based Kriging , 2015, 1502.03939.

[17]  A. O'Hagan,et al.  Bayesian calibration of computer models , 2001 .

[18]  H. Najm,et al.  A stochastic projection method for fluid flow II.: random process , 2002 .

[19]  G. Gary Wang,et al.  Survey of modeling and optimization strategies to solve high-dimensional design problems with computationally-expensive black-box functions , 2010 .

[20]  Carl E. Rasmussen,et al.  Gaussian Processes for Machine Learning (GPML) Toolbox , 2010, J. Mach. Learn. Res..

[21]  Evgueni A. Haroutunian,et al.  Information Theory and Statistics , 2011, International Encyclopedia of Statistical Science.

[22]  Qiqi Wang,et al.  Residual Minimizing Model Interpolation for Parameterized Nonlinear Dynamical Systems , 2010, SIAM J. Sci. Comput..

[23]  Bertrand Iooss,et al.  An efficient methodology for modeling complex computer codes with Gaussian processes , 2008, Comput. Stat. Data Anal..

[24]  Gerardo Rubino,et al.  Rare Event Simulation using Monte Carlo Methods , 2009 .

[25]  Dongbin Xiu,et al.  The Wiener-Askey Polynomial Chaos for Stochastic Differential Equations , 2002, SIAM J. Sci. Comput..

[26]  W. F. Kibble An extension of a theorem of Mehler's on Hermite polynomials , 1945, Mathematical Proceedings of the Cambridge Philosophical Society.

[27]  Bin Liang,et al.  ERROR AND UNCERTAINTY QUANTIFICATION AND SENSITIVITY ANALYSIS IN MECHANICS COMPUTATIONAL MODELS , 2011 .

[28]  Michael D. Shields,et al.  Surrogate-enhanced stochastic search algorithms to identify implicitly defined functions for reliability analysis , 2016 .

[29]  Nicolas Gayton,et al.  AK-MCS: An active learning reliability method combining Kriging and Monte Carlo Simulation , 2011 .

[30]  A. O'Hagan,et al.  Probabilistic sensitivity analysis of complex models: a Bayesian approach , 2004 .

[31]  Alexander J. Smola,et al.  Support Vector Regression Machines , 1996, NIPS.

[32]  Stefano Marelli,et al.  Rare Event Estimation Using Polynomial-Chaos Kriging , 2017 .

[33]  R. Ghanem,et al.  Stochastic Finite-Element Analysis of Seismic Soil-Structure Interaction , 2002 .

[34]  Sonja Kuhnt,et al.  Design and analysis of computer experiments , 2010 .

[35]  J. Freidman,et al.  Multivariate adaptive regression splines , 1991 .

[36]  Bruno Sudret,et al.  PC-Kriging: A new meta-modelling method and its application to quantile estimation , 2014 .

[37]  Bruno Sudret,et al.  Combining polynomial chaos expansions and Kriging for solving structural reliability problems , 2014 .

[38]  Edoardo Patelli,et al.  Robust artificial neural network for reliability and sensitivity analyses of complex non-linear systems , 2017, Neural Networks.

[39]  Aleksey V. Nenarokomov,et al.  Uncertainties in parameter estimation: the inverse problem , 1995 .