Offspring population size matters when comparing evolutionary algorithms with self-adjusting mutation rates

We analyze the performance of the 2-rate (1 + λ) Evolutionary Algorithm (EA) with self-adjusting mutation rate control, its 3-rate counterpart, and a (1 + λ) EA variant using multiplicative update rules on the OneMax problem. We compare their efficiency for offspring population sizes ranging up to λ = 3, 200 and problem sizes up to n = 100,000. Our empirical results show that the ranking of the algorithms is very consistent across all tested dimensions, but strongly depends on the population size. While for small values of λ the 2-rate EA performs best, the multiplicative updates become superior for starting for some threshold value of λ between 50 and 100. Interestingly, for population sizes around 50, the (1 + λ) EA with static mutation rates performs on par with the best of the self-adjusting algorithms. We also consider how the lower bound pmin for the mutation rate influences the efficiency of the algorithms. We observe that for the 2-rate EA and the EA with multiplicative update rules the more generous bound pmin = 1/n2 gives better results than pmin = 1/n when λ is small. For both algorithms the situation reverses for large λ.

[1]  Thomas Bäck,et al.  The Interaction of Mutation Rate, Selection, and Self-Adaptation Within a Genetic Algorithm , 1992, PPSN.

[2]  Dirk Sudholt,et al.  Adaptive population models for offspring populations and parallel evolutionary algorithms , 2011, FOGA '11.

[3]  Frédéric Saubion,et al.  Autonomous operator management for evolutionary algorithms , 2010, J. Heuristics.

[4]  Thomas Jansen,et al.  Analyzing Evolutionary Algorithms , 2015, Natural Computing Series.

[5]  Hao Wang,et al.  Towards a theory-guided benchmarking suite for discrete black-box optimization heuristics: profiling (1 + λ) EA variants on onemax and leadingones , 2018, GECCO.

[6]  Carola Doerr,et al.  Complexity Theory for Discrete Black-Box Optimization Heuristics , 2018, Theory of Evolutionary Computation.

[7]  Thomas Weise,et al.  Difficult features of combinatorial optimization problems and the tunable w-model benchmark problem for simulating them , 2018, GECCO.

[8]  Johannes Lengler,et al.  A General Dichotomy of Evolutionary Algorithms on Monotone Functions , 2018, IEEE Transactions on Evolutionary Computation.

[9]  Michel Gendreau,et al.  Hyper-heuristics: a survey of the state of the art , 2013, J. Oper. Res. Soc..

[10]  Thomas Jansen,et al.  Analyzing Evolutionary Algorithms: The Computer Science Perspective , 2012 .

[11]  Benjamin Doerr,et al.  Self-Adjusting Mutation Rates with Provably Optimal Success Rules , 2019, Algorithmica.

[12]  Anne Auger,et al.  Theory of Randomized Search Heuristics: Foundations and Recent Developments , 2011, Theory of Randomized Search Heuristics.

[13]  Michèle Sebag,et al.  Analyzing bandit-based adaptive operator selection mechanisms , 2010, Annals of Mathematics and Artificial Intelligence.

[14]  Thomas Jansen,et al.  Performance analysis of randomised search heuristics operating with a fixed budget , 2014, Theor. Comput. Sci..

[15]  Michèle Sebag,et al.  Extreme Value Based Adaptive Operator Selection , 2008, PPSN.

[16]  Anne Auger,et al.  Theory of Randomized Search Heuristics , 2012, Algorithmica.

[17]  Benjamin Doerr,et al.  Optimal Static and Self-Adjusting Parameter Choices for the (1+(λ,λ))\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$( , 2017, Algorithmica.

[18]  Dirk Thierens On benchmark properties for adaptive operator selection , 2009, GECCO '09.

[19]  Carola Doerr,et al.  Towards a More Practice-Aware Runtime Analysis of Evolutionary Algorithms , 2017, ArXiv.

[20]  Benjamin Doerr,et al.  Optimal Parameter Choices via Precise Black-Box Analysis , 2016, GECCO.

[21]  Benjamin Doerr,et al.  From black-box complexity to designing new genetic algorithms , 2015, Theor. Comput. Sci..

[22]  Kenneth A. De Jong,et al.  Design and Management of Complex Technical Processes and Systems by Means of Computational Intelligence Methods on the Choice of the Offspring Population Size in Evolutionary Algorithms on the Choice of the Offspring Population Size in Evolutionary Algorithms , 2004 .

[23]  Mark Hoogendoorn,et al.  Parameter Control in Evolutionary Algorithms: Trends and Challenges , 2015, IEEE Transactions on Evolutionary Computation.

[24]  Petros Koumoutsakos,et al.  Learning Probability Distributions in Continuous Evolutionary Algorithms - a Comparative Review , 2004, Nat. Comput..

[25]  Markus Wagner,et al.  On the Effectiveness of Simple Success-Based Parameter Selection Mechanisms for Two Classical Discrete Black-Box Optimization Benchmark Problems , 2018, ArXiv.

[26]  Benjamin Doerr,et al.  The (1+λ) evolutionary algorithm with self-adjusting mutation rate , 2017, GECCO.

[27]  Irene Moser,et al.  A Systematic Literature Review of Adaptive Parameter Control Methods for Evolutionary Algorithms , 2016, ACM Comput. Surv..

[28]  Zbigniew Michalewicz,et al.  Parameter Control in Evolutionary Algorithms , 2007, Parameter Setting in Evolutionary Algorithms.

[29]  Benjamin Doerr,et al.  The ($$1+\lambda $$1+λ) Evolutionary Algorithm with Self-Adjusting Mutation Rate , 2018, Algorithmica.

[30]  Per Kristian Lehre,et al.  Unbiased Black-Box Complexity of Parallel Search , 2014, PPSN.

[31]  Markus Wagner,et al.  Simple on-the-fly parameter selection mechanisms for two classical discrete black-box optimization benchmark problems , 2018, GECCO.

[32]  Thomas Bäck,et al.  Optimal Mutation Rates in Genetic Search , 1993, ICGA.

[33]  Carola Doerr,et al.  A Simple Proof for the Usefulness of Crossover in Black-Box Optimization , 2018, PPSN.