Explorative data analysis of time series based algorithm features of CMA-ES variants

In this study, we analyze behaviours of the well-known CMA-ES by extracting the time-series features on its dynamic strategy parameters. An extensive experiment was conducted on twelve CMA-ES variants and 24 test problems taken from the BBOB (Black-Box Optimization Bench-marking) testbed, where we used two different cutoff times to stop those variants. We utilized the tsfresh package for extracting the features and performed the feature selection procedure using the Boruta algorithm, resulting in 32 features to distinguish either CMA-ES variants or the problems. After measuring the number of predefined targets reached by those variants, we contrive to predict those measured values on each test problem using the feature. From our analysis, we saw that the features can classify the CMA-ES variants, or the function groups decently, and show a potential for predicting the performance of those variants. We conducted a hierarchical clustering analysis on the test problems and noticed a drastic change in the clustering outcome when comparing the longer cutoff time to the shorter one, indicating a huge change in search behaviour of the algorithm. In general, we found that with longer time series, the predictive power of the time series features increase.

[1]  Hao Wang,et al.  Evolving the structure of Evolution Strategies , 2016, 2016 IEEE Symposium Series on Computational Intelligence (SSCI).

[2]  P. A. Blight The Analysis of Time Series: An Introduction , 1991 .

[3]  Nikolaus Hansen,et al.  Completely Derandomized Self-Adaptation in Evolution Strategies , 2001, Evolutionary Computation.

[4]  Hao Wang,et al.  Algorithm configuration data mining for CMA evolution strategies , 2017, GECCO.

[5]  Kumudha Raimond,et al.  A Survey on Optimization Algorithms for Optimizing the Numerical Functions , 2013 .

[6]  Anne Auger,et al.  Real-Parameter Black-Box Optimization Benchmarking 2009: Noiseless Functions Definitions , 2009 .

[7]  Antonio Bolufé Röhler,et al.  Evolution strategies with thresheld convergence , 2015, 2015 IEEE Congress on Evolutionary Computation (CEC).

[8]  Dirk V. Arnold,et al.  Improving Evolution Strategies through Active Covariance Matrix Adaptation , 2006, 2006 IEEE International Conference on Evolutionary Computation.

[9]  Bernd Bischl,et al.  Exploratory landscape analysis , 2011, GECCO '11.

[10]  Anne Auger,et al.  Mirrored Sampling and Sequential Selection for Evolution Strategies , 2010, PPSN.

[11]  Hao Wang,et al.  Mirrored orthogonal sampling with pairwise selection in evolution strategies , 2014, SAC.

[12]  Witold R. Rudnicki,et al.  Feature Selection with the Boruta Package , 2010 .

[13]  Gaël Varoquaux,et al.  Scikit-learn: Machine Learning in Python , 2011, J. Mach. Learn. Res..

[14]  Olivier Teytaud,et al.  Quasi-random Numbers Improve the CMA-ES on the BBOB Testbed , 2015, Artificial Evolution.

[15]  Nikolaus Hansen,et al.  CMA-ES with Two-Point Step-Size Adaptation , 2008, ArXiv.

[16]  S. Csiszar Optimization Algorithms (survey and analysis) , 2007, 2007 International Symposium on Logistics and Industrial Informatics.

[17]  Geoffrey E. Hinton,et al.  Visualizing Data using t-SNE , 2008 .

[18]  Nikolaus Hansen,et al.  The CMA Evolution Strategy: A Tutorial , 2016, ArXiv.

[19]  Andreas W. Kempa-Liehr,et al.  Time Series FeatuRe Extraction on basis of Scalable Hypothesis tests (tsfresh - A Python package) , 2018, Neurocomputing.

[20]  R. Geoff Dromey,et al.  An algorithm for the selection problem , 1986, Softw. Pract. Exp..

[21]  Bernd Bischl,et al.  Analyzing the BBOB Results by Means of Benchmarking Concepts , 2015, Evolutionary Computation.

[22]  Andreas W. Kempa-Liehr,et al.  Distributed and parallel time series feature extraction for industrial big data applications , 2016, ArXiv.

[23]  Anne Auger,et al.  Mirrored sampling in evolution strategies with weighted recombination , 2011, GECCO '11.

[24]  Nikolaus Hansen,et al.  Benchmarking a BI-population CMA-ES on the BBOB-2009 function testbed , 2009, GECCO '09.

[25]  Anne Auger,et al.  A median success rule for non-elitist evolution strategies: study of feasibility , 2013, GECCO '13.

[26]  Pascal Kerschke,et al.  Comprehensive Feature-Based Landscape Analysis of Continuous and Constrained Optimization Problems Using the R-Package Flacco , 2017, Studies in Classification, Data Analysis, and Knowledge Organization.

[27]  Pierre Geurts,et al.  Extremely randomized trees , 2006, Machine Learning.