Additive Kernels for Gaussian Process Modeling

Gaussian Process (GP) models are often used as mathematical approximations of computationally expensive experiments. Provided that its kernel is suitably chosen and that enough data is available to obtain a reasonable fit of the simulator, a GP model can beneficially be used for tasks such as prediction, optimization, or Monte-Carlo-based quantification of uncertainty. However, the former conditions become unrealistic when using classical GPs as the dimension of input increases. One popular alternative is then to turn to Generalized Additive Models (GAMs), relying on the assumption that the simulator's response can approximately be decomposed as a sum of univariate functions. If such an approach has been successfully applied in approximation, it is nevertheless not completely compatible with the GP framework and its versatile applications. The ambition of the present work is to give an insight into the use of GPs for additive models by integrating additivity within the kernel, and proposing a parsimonious numerical method for data-driven parameter estimation. The first part of this article deals with the kernels naturally associated to additive processes and the properties of the GP models based on such kernels. The second part is dedicated to a numerical procedure based on relaxation for additive kernel parameter estimation. Finally, the efficiency of the proposed method is illustrated and compared to other approaches on Sobol's g-function.

[1]  R. Tibshirani,et al.  Linear Smoothers and Additive Models , 1989 .

[2]  R. Tibshirani,et al.  Generalized Additive Models , 1991 .

[3]  Joseph A. C. Delaney Sensitivity analysis , 2018, The African Continental Free Trade Area: Economic and Distributional Effects.

[4]  Olivier Roustant,et al.  Data-driven Kriging models based on FANOVA-decomposition , 2012, Stat. Comput..

[5]  Carl E. Rasmussen,et al.  Gaussian processes for machine learning , 2005, Adaptive computation and machine learning.

[6]  Henry P. Wynn,et al.  Screening, predicting, and computer experiments , 1992 .

[7]  W. Newey,et al.  Kernel Estimation of Partial Means and a General Variance Estimator , 1994, Econometric Theory.

[8]  David K. Smith,et al.  Mathematical Programming: Theory and Algorithms , 1986 .

[9]  A. OHagan,et al.  Bayesian analysis of computer code outputs: A tutorial , 2006, Reliab. Eng. Syst. Saf..

[10]  David Ginsbourger,et al.  A note on the choice and the estimation of Kriging models for the analysis of deterministic computer experiments , 2009 .

[11]  Thomas J. Santner,et al.  The Design and Analysis of Computer Experiments , 2003, Springer Series in Statistics.

[12]  Mike Rees,et al.  5. Statistics for Spatial Data , 1993 .

[13]  Bertrand Iooss,et al.  An efficient methodology for modeling complex computer codes with Gaussian processes , 2008, Comput. Stat. Data Anal..

[14]  C. J. Stone,et al.  Additive Regression and Other Nonparametric Models , 1985 .

[15]  Runze Li,et al.  Design and Modeling for Computer Experiments , 2005 .

[16]  Michel Minoux,et al.  Mathematical Programming , 1986 .

[17]  Noel A Cressie,et al.  Statistics for Spatial Data. , 1992 .

[18]  Rhodri Hayward,et al.  Screening , 2008, The Lancet.