Small ensembles of kriging models for optimization

The Efficient Global Optimization (EGO) algorithm uses a conditional Gaus-sian Process (GP) to approximate an objective function known at a finite number of observation points and sequentially adds new points which maximize the Expected Improvement criterion according to the GP. The important factor that controls the efficiency of EGO is the GP covariance function (or kernel) which should be chosen according to the objective function. Traditionally, a pa-rameterized family of covariance functions is considered whose parameters are learned through statistical procedures such as maximum likelihood or cross-validation. However, it may be questioned whether statistical procedures for learning covariance functions are the most efficient for optimization as they target a global agreement between the GP and the observations which is not the ultimate goal of optimization. Furthermore, statistical learning procedures are computationally expensive. The main alternative to the statistical learning of the GP is self-adaptation, where the algorithm tunes the kernel parameters based on their contribution to objective function improvement. After questioning the possibility of self-adaptation for kriging based optimizers, this paper proposes a novel approach for tuning the length-scale of the GP in EGO: At each iteration, a small ensemble of kriging models structured by their length-scales is created. All of the models contribute to an iterate in an EGO-like fashion. Then, the set of models is densified around the model whose length-scale yielded the best iterate and further points are produced. Numerical experiments are provided which motivate the use of many length-scales. The tested implementation does not perform better than the classical EGO algorithm in a sequential context but show the potential of the approach for parallel implementations.

[1]  Kevin Leyton-Brown,et al.  An evaluation of sequential model-based optimization for expensive blackbox functions , 2013, GECCO.

[2]  Layne T. Watson,et al.  Efficient global optimization algorithm assisted by multiple surrogate techniques , 2012, Journal of Global Optimization.

[3]  Donald R. Jones,et al.  Global optimization of deceptive functions with sparse sampling , 2008 .

[4]  Jack P. C. Kleijnen,et al.  Simulation-optimization via Kriging and bootstrapping: a survey , 2014, J. Simulation.

[5]  Marcus R. Frean,et al.  Using Gaussian Processes to Optimize Expensive Functions , 2008, Australasian Conference on Artificial Intelligence.

[6]  Christopher K. I. Williams,et al.  Gaussian Processes for Machine Learning (Adaptive Computation and Machine Learning) , 2005 .

[7]  Thomas Stützle,et al.  GECCO 2007: Genetic and Evolutionary Computation Conference , 2007 .

[8]  Mark Harman,et al.  Genetic and Evolutionary Computation (GECCO) , 2003 .

[9]  Eric Walter,et al.  An informational approach to the global optimization of expensive-to-evaluate functions , 2006, J. Glob. Optim..

[10]  Masoud Rais-Rohani,et al.  Ensemble of Metamodels with Optimized Weight Factors , 2008 .

[11]  Hao Zhang,et al.  Kriging and cross‐validation for massive spatial data , 2009 .

[12]  Rodolphe Le Riche,et al.  A detailed analysis of kernel parameters in Gaussian process-based optimization , 2015 .

[13]  Roberto Battiti,et al.  Learning and Intelligent Optimization , 2017, Lecture Notes in Computer Science.

[14]  Nando de Freitas,et al.  Bayesian Optimization in High Dimensions via Random Embeddings , 2013, IJCAI.

[15]  Julien Bect,et al.  Robust Gaussian Process-Based Global Optimization Using a Fully Bayesian Expected Improvement Criterion , 2011, LION.

[16]  Bin Li,et al.  An evolution strategy assisted by an ensemble of local Gaussian process models , 2013, GECCO '13.

[17]  Thomas Bäck,et al.  Evolutionary algorithms in theory and practice - evolution strategies, evolutionary programming, genetic algorithms , 1996 .

[18]  Anirban Chaudhuri,et al.  Estimating Feasibility Using Multiple Surrogates and ROC Curves , 2013 .

[19]  Donald R. Jones,et al.  A Taxonomy of Global Optimization Methods Based on Response Surfaces , 2001, J. Glob. Optim..

[20]  Enrique del Castillo,et al.  A Bayesian Approach to Sequential Optimization based on Computer Experiments , 2015, Qual. Reliab. Eng. Int..

[21]  Carl E. Rasmussen,et al.  Gaussian processes for machine learning , 2005, Adaptive computation and machine learning.

[22]  Zhiliang Ying,et al.  Asymptotic properties of a maximum likelihood estimator with data from a Gaussian process , 1991 .

[23]  Runze Li,et al.  Analysis of Computer Experiments Using Penalized Likelihood in Gaussian Kriging Models , 2005, Technometrics.

[24]  R. Haftka,et al.  Ensemble of surrogates , 2007 .

[25]  François Bachoc,et al.  Cross Validation and Maximum Likelihood estimations of hyper-parameters of Gaussian processes with model misspecification , 2013, Comput. Stat. Data Anal..

[26]  Donald R. Jones,et al.  Efficient Global Optimization of Expensive Black-Box Functions , 1998, J. Glob. Optim..

[27]  Bernd Bischl,et al.  PROGRESS: Progressive Reinforcement-Learning-Based Surrogate Selection , 2013, LION.