Interpolating Local and Global Search by Controlling the Variance of Standard Bit Mutation

A key property underlying the success of evolutionary algorithms (EAs) is their global search behavior, which allows the algorithms to "jump" from a current state to other parts of the search space, thereby avoiding to get stuck in local optima. This property is obtained through a random choice of the radius at which offspring are sampled from previously evaluated solutions. It is well known that, thanks to this global search behavior, the probability that an EA using standard bit mutation finds a global optimum of an arbitrary function f : {0, 1}n → ℝ tends to one as the number of function evaluations grows. This advantage over heuristics using a fixed search radius, however, comes at the cost of using non-optimal step sizes also in those regimes in which the optimal rate is stable for a long time. This downside results in significant performance losses for many standard benchmark problems.We introduce in this work a simple way to interpolate between the random global search of EAs and their deterministic counterparts which sample from a fixed radius only. To this end, we introduce normalized standard bit mutation, in which the binomial choice of the search radius is replaced by a normal distribution. Normalized standard bit mutation allows a straightforward way to control its variance, and hence the degree of randomness involved. We experiment with a self-adjusting choice of this variance, and demonstrate its effectiveness for the two classic benchmark problems LeadingOnes and OneMax. Our work thereby also touches a largely ignored question in discrete evolutionary computation: multi-dimensional parameter control.

[1]  Thomas Jansen,et al.  Analysis of evolutionary algorithms: from computational complexity analysis to algorithm engineering , 2011, FOGA '11.

[2]  Markus Wagner,et al.  On the Effectiveness of Simple Success-Based Parameter Selection Mechanisms for Two Classical Discrete Black-Box Optimization Benchmark Problems , 2018, ArXiv.

[3]  Benjamin Doerr,et al.  Optimal Parameter Choices via Precise Black-Box Analysis , 2016, GECCO.

[4]  Johannes Lengler,et al.  A General Dichotomy of Evolutionary Algorithms on Monotone Functions , 2018, IEEE Transactions on Evolutionary Computation.

[5]  Carola Doerr,et al.  Maximizing Drift Is Not Optimal for Solving OneMax , 2019, Evolutionary Computation.

[6]  Carsten Witt,et al.  The Interplay of Population Size and Mutation Probability in the (1+λ) EA on OneMax , 2017, Algorithmica.

[7]  Hao Wang,et al.  Towards a theory-guided benchmarking suite for discrete black-box optimization heuristics: profiling (1 + λ) EA variants on onemax and leadingones , 2018, GECCO.

[8]  Benjamin Doerr,et al.  Better Runtime Guarantees via Stochastic Domination , 2018, EvoCOP.

[9]  Carsten Witt,et al.  The Interplay of Population Size and Mutation Probability in the ($$1+\lambda $$1+λ) EA on OneMax , 2016, Algorithmica.

[10]  Hao Wang,et al.  IOHprofiler: A Benchmarking and Profiling Tool for Iterative Optimization Heuristics , 2018, ArXiv.

[11]  Markus Wagner,et al.  Heavy-Tailed Mutation Operators in Single-Objective Combinatorial Optimization , 2018, PPSN.

[12]  Kenneth A. De Jong,et al.  Design and Management of Complex Technical Processes and Systems by Means of Computational Intelligence Methods on the Choice of the Offspring Population Size in Evolutionary Algorithms on the Choice of the Offspring Population Size in Evolutionary Algorithms , 2004 .

[13]  Mark Hoogendoorn,et al.  Parameter Control in Evolutionary Algorithms: Trends and Challenges , 2015, IEEE Transactions on Evolutionary Computation.

[14]  Benjamin Doerr,et al.  The (1+λ) evolutionary algorithm with self-adjusting mutation rate , 2017, GECCO.

[15]  Maxim Buzdalov,et al.  Evaluation of heavy-tailed mutation operator on maximum flow test generation problem , 2017, GECCO.

[16]  Markus Wagner,et al.  Escaping large deceptive basins of attraction with heavy-tailed mutation operators , 2018, GECCO.

[17]  Carola Doerr,et al.  Dynamic parameter choices in evolutionary computation , 2018, GECCO.

[18]  Benjamin Doerr,et al.  Fast genetic algorithms , 2017, GECCO.

[19]  Carola Doerr,et al.  A Simple Proof for the Usefulness of Crossover in Black-Box Optimization , 2018, PPSN.

[20]  Pietro Simone Oliveto,et al.  Fast Artificial Immune Systems , 2018, PPSN.

[21]  Pietro Simone Oliveto,et al.  On the runtime analysis of generalised selection hyper-heuristics for pseudo-boolean optimisation , 2017, GECCO.