Integrated vs. sequential approaches for selecting and tuning CMA-ES variants

When faced with a specific optimization problem, deciding which algorithm to apply is always a difficult task. Not only is there a vast variety of algorithms to select from, but these algorithms are often controlled by many hyperparameters, which need to be suitably tuned in order to achieve peak performance. Usually, the problem of selecting and configuring the optimization algorithm is addressed sequentially, by first selecting a suitable algorithm and then tuning it for the application at hand. Integrated approaches, commonly known as Combined Algorithm Selection and Hyperparameter (CASH) solvers, have shown promise in several applications. In this work we compare sequential and integrated approaches for selecting and tuning the best out of the 4,608 variants of the modular Covariance Matrix Adaptation Evolution Strategy (CMA-ES). We show that the ranking of these variants depends to a large extent on the quality of the hyperparameters. Sequential approaches are therefore likely to recommend sub-optimal choices. Integrated approaches, in contrast, manage to provide competitive results at much smaller computational cost. We also highlight important differences in the search behavior of two CASH approaches, which build on racing (irace) and on model-based optimization (MIP-EGO), respectively.

[1]  Hao Wang,et al.  Evolving the structure of Evolution Strategies , 2016, 2016 IEEE Symposium Series on Computational Intelligence (SSCI).

[2]  Heike Trautmann,et al.  Automated Algorithm Selection: Survey and Perspectives , 2018, Evolutionary Computation.

[3]  Thomas Bäck,et al.  Contemporary Evolution Strategies , 2013, Natural Computing Series.

[4]  Kevin Leyton-Brown,et al.  Sequential Model-Based Optimization for General Algorithm Configuration , 2011, LION.

[5]  Aaron Klein,et al.  Efficient and Robust Automated Machine Learning , 2015, NIPS.

[6]  Leslie Pérez Cáceres,et al.  The irace package: Iterated racing for automatic algorithm configuration , 2016 .

[7]  John R. Rice,et al.  The Algorithm Selection Problem , 1976, Adv. Comput..

[8]  Anne Auger,et al.  Comparing results of 31 algorithms from the black-box optimization benchmarking BBOB-2009 , 2010, GECCO '10.

[9]  Nikolaus Hansen,et al.  Adapting arbitrary normal mutation distributions in evolution strategies: the covariance matrix adaptation , 1996, Proceedings of IEEE International Conference on Evolutionary Computation.

[10]  Bernd Bischl,et al.  Exploratory landscape analysis , 2011, GECCO '11.

[11]  Nikolaus Hansen,et al.  Benchmarking a BI-population CMA-ES on the BBOB-2009 function testbed , 2009, GECCO '09.

[12]  Leyton-BrownKevin,et al.  Auto-WEKA 2.0 , 2017 .

[13]  Zbigniew Michalewicz,et al.  Parameter control in evolutionary algorithms , 1999, IEEE Trans. Evol. Comput..

[14]  Anna Syberfeldt,et al.  Parameter tuned CMA-ES on the CEC'15 expensive problems , 2015, 2015 IEEE Congress on Evolutionary Computation (CEC).

[15]  Anne Auger,et al.  Mirrored sampling in evolution strategies with weighted recombination , 2011, GECCO '11.

[16]  Thomas Bäck,et al.  Online selection of CMA-ES variants , 2019, GECCO.

[17]  Dirk V. Arnold,et al.  Improving Evolution Strategies through Active Covariance Matrix Adaptation , 2006, 2006 IEEE International Conference on Evolutionary Computation.

[18]  Heike Trautmann,et al.  The R-Package FLACCO for exploratory landscape analysis with applications to multi-objective optimization problems , 2016, 2016 IEEE Congress on Evolutionary Computation (CEC).

[19]  Anne Auger,et al.  Mirrored Sampling and Sequential Selection for Evolution Strategies , 2010, PPSN.

[20]  Olivier Teytaud,et al.  Algorithms (X, sigma, eta): Quasi-random Mutations for Evolution Strategies , 2005, Artificial Evolution.

[21]  Antonio Bolufé Röhler,et al.  Evolution strategies with thresheld convergence , 2015, 2015 IEEE Congress on Evolutionary Computation (CEC).

[22]  Carola Doerr,et al.  Adaptive landscape analysis , 2019, GECCO.

[23]  Martin Holena,et al.  Landscape analysis of gaussian process surrogates for the covariance matrix adaptation evolution strategy , 2019, GECCO.

[24]  Leslie Pérez Cáceres,et al.  An Experimental Study of Adaptive Capping in irace , 2017, LION.

[25]  Johann Dréo,et al.  Expressiveness and robustness of landscape features , 2019, GECCO.

[26]  Thomas Bäck,et al.  Towards an Adaptive CMA-ES Configurator , 2018, PPSN.

[27]  Hao Wang,et al.  A new acquisition function for Bayesian optimization based on the moment-generating function , 2017, 2017 IEEE International Conference on Systems, Man, and Cybernetics (SMC).

[28]  Hao Wang,et al.  Algorithm configuration data mining for CMA evolution strategies , 2017, GECCO.

[29]  Hao Wang,et al.  Cooling Strategies for the Moment-Generating Function in Bayesian Global Optimization , 2018, 2018 IEEE Congress on Evolutionary Computation (CEC).

[30]  Kevin Leyton-Brown,et al.  Auto-WEKA: combined selection and hyperparameter optimization of classification algorithms , 2012, KDD.

[31]  Marc Schoenauer,et al.  Per instance algorithm configuration of CMA-ES with limited budget , 2017, GECCO.

[32]  Hao Wang,et al.  Mirrored orthogonal sampling with pairwise selection in evolution strategies , 2014, SAC.

[33]  David H. Wolpert,et al.  No free lunch theorems for optimization , 1997, IEEE Trans. Evol. Comput..

[34]  Johannes Lengler,et al.  A General Dichotomy of Evolutionary Algorithms on Monotone Functions , 2018, IEEE Transactions on Evolutionary Computation.

[35]  Nikolaus Hansen,et al.  A restart CMA evolution strategy with increasing population size , 2005, 2005 IEEE Congress on Evolutionary Computation.