Self-Adjusting Evolutionary Algorithms for Multimodal Optimization

Recent theoretical research has shown that self-adjusting and self-adaptive mechanisms can provably outperform static settings in evolutionary algorithms for binary search spaces. However, the vast majority of these studies focuses on unimodal functions which do not require the algorithm to flip several bits simultaneously to make progress. In fact, existing self-adjusting algorithms are not designed to detect local optima and do not have any obvious benefit to cross large Hamming gaps. We suggest a mechanism called stagnation detection that can be added as a module to existing evolutionary algorithms (both with and without prior self-adjusting schemes). Added to a simple (1 + 1) EA, we prove an expected runtime on the well-known Jump benchmark that corresponds to an asymptotically optimal parameter setting and outperforms other mechanisms for multimodal optimization like heavy-tailed mutation. We also investigate the module in the context of a self-adjusting (1 + λ) EA and show that it combines the previous benefits of this algorithm on unimodal problems with more efficient multimodal optimization. To explore the limitations of the approach, we additionally present an example where both self-adjusting mechanisms, including stagnation detection, do not help to find a beneficial setting of the mutation rate. Finally, we investigate our module for stagnation detection experimentally.

[1]  Carsten Witt,et al.  Population size versus runtime of a simple evolutionary algorithm , 2008, Theor. Comput. Sci..

[2]  Benjamin Doerr,et al.  The unrestricted black-box complexity of jump functions , 2017, GECCO.

[3]  Thomas Jansen,et al.  On the analysis of the (1+1) evolutionary algorithm , 2002, Theor. Comput. Sci..

[4]  Benjamin Doerr,et al.  Fast genetic algorithms , 2017, GECCO.

[5]  Thomas Bäck,et al.  Interpolating Local and Global Search by Controlling the Variance of Standard Bit Mutation , 2019, 2019 IEEE Congress on Evolutionary Computation (CEC).

[6]  Pierre Hansen,et al.  Variable Neighborhood Search , 2018, Handbook of Heuristics.

[7]  Markus Wagner,et al.  Escaping large deceptive basins of attraction with heavy-tailed mutation operators , 2018, GECCO.

[8]  C. Witt Population size vs. runtime of a simple EA , 2003, The 2003 Congress on Evolutionary Computation, 2003. CEC '03..

[9]  Thomas Jansen,et al.  Exploring the Explorative Advantage of the Cooperative Coevolutionary (1+1) EA , 2003, GECCO.

[10]  Mario Alejandro,et al.  An empirical evaluation of success-based parameter control mechanisms for evolutionary algorithms , 2019, GECCO.

[11]  Johannes Lengler,et al.  A General Dichotomy of Evolutionary Algorithms on Monotone Functions , 2018, IEEE Transactions on Evolutionary Computation.

[12]  Anirban Mukhopadhyay,et al.  Exploration and Exploitation Without Mutation: Solving the Jump Function in \varTheta (n) Time , 2018, PPSN.

[13]  Mahmoud Fouz,et al.  Quasirandom evolutionary algorithms , 2010, GECCO '10.

[14]  Benjamin Doerr,et al.  Optimal Static and Self-Adjusting Parameter Choices for the (1+(λ,λ))\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$( , 2017, Algorithmica.

[15]  Arina Buzdalova,et al.  Offspring population size matters when comparing evolutionary algorithms with self-adjusting mutation rates , 2019, GECCO.

[16]  Benjamin Doerr,et al.  Theory of Parameter Control for Discrete Black-Box Optimization: Provable Performance Gains Through Dynamic Parameter Choices , 2018, Theory of Evolutionary Computation.

[17]  Dirk Sudholt,et al.  Adaptive population models for offspring populations and parallel evolutionary algorithms , 2011, FOGA '11.

[18]  Hao Wang,et al.  Towards a theory-guided benchmarking suite for discrete black-box optimization heuristics: profiling (1 + λ) EA variants on onemax and leadingones , 2018, GECCO.

[19]  Pietro Simone Oliveto,et al.  Fast Artificial Immune Systems , 2018, PPSN.

[20]  Benjamin Doerr,et al.  Significance-Based Estimation-of-Distribution Algorithms , 2018, IEEE Transactions on Evolutionary Computation.

[21]  Carsten Witt,et al.  Tight Bounds on the Optimization Time of a Randomized Search Heuristic on Linear Functions† , 2013, Combinatorics, Probability and Computing.

[22]  Carsten Witt,et al.  Runtime Analysis of the ( + 1) EA on Simple Pseudo-Boolean Functions , 2006, Evolutionary Computation.

[23]  Benjamin Doerr,et al.  A tight runtime analysis for the cGA on jump functions: EDAs can cross fitness valleys at no extra cost , 2019, GECCO.

[24]  Benjamin Doerr,et al.  Static and Self-Adjusting Mutation Strengths for Multi-valued Decision Variables , 2018, Algorithmica.

[25]  Kenneth A. De Jong,et al.  Design and Management of Complex Technical Processes and Systems by Means of Computational Intelligence Methods on the Choice of the Offspring Population Size in Evolutionary Algorithms on the Choice of the Offspring Population Size in Evolutionary Algorithms , 2004 .

[26]  Ingo Wegener,et al.  Methods for the Analysis of Evolutionary Algorithms on Pseudo-Boolean Functions , 2003 .

[27]  Benjamin Doerr,et al.  Runtime Analysis for Self-adaptive Mutation Rates , 2018, Algorithmica.

[28]  Duc-Cuong Dang,et al.  Self-adaptation of Mutation Rates in Non-elitist Populations , 2016, PPSN.

[29]  Aishwaryaprajna,et al.  The benefits and limitations of voting mechanisms in evolutionary optimisation , 2019, FOGA '19.

[30]  Benjamin Doerr,et al.  The (1+λ) evolutionary algorithm with self-adjusting mutation rate , 2017, GECCO.

[31]  Andrei Lissovoi,et al.  Simple Hyper-Heuristics Control the Neighbourhood Size of Randomised Local Search Optimally for LeadingOnes* , 2018, Evolutionary Computation.

[32]  Markus Wagner,et al.  Sensitivity of Parameter Control Mechanisms with Respect to Their Initialization , 2018, PPSN.

[33]  Elena Marchiori,et al.  Evolutionary Algorithms with On-the-Fly Population Size Adjustment , 2004, PPSN.

[34]  Benjamin Doerr,et al.  The ($$1+\lambda $$1+λ) Evolutionary Algorithm with Self-Adjusting Mutation Rate , 2018, Algorithmica.

[35]  B. Hajek Hitting-time and occupation-time bounds implied by drift analysis with applications , 1982, Advances in Applied Probability.

[36]  Benjamin Doerr,et al.  A tight runtime analysis for the (1 + (λ, λ)) GA on leadingones , 2019, FOGA '19.

[37]  Per Kristian Lehre,et al.  Dynamic evolutionary optimisation: an analysis of frequency and magnitude of change , 2009, GECCO.

[38]  Benjamin Doerr,et al.  The (1 + (λ,λ)) GA is even faster on multimodal problems , 2020, GECCO.

[39]  Carsten Witt,et al.  Runtime Analysis of the ( μ +1) EA on Simple Pseudo-Boolean Functions , 2006 .