The Impact of Parametrization on Randomized Search Heuristics

The Impact of Parametrization on Randomized Search Heuristics In this work we present runtime analyses of randomized search heuristics (RSH) in various settings that are determined by parameters of the problems, the algorithms and also exogenous parameters like noise. In the process we provide new techniques for the theoretical analysis of RSH as well as new optimization algorithms. We consider the following topics. Escaping local optima using local search. We analyze memetic algorithms, i.e. evolutionary algorithms equipped with a local search after mutation. To this end we consider the (1+1) EA equipped with Standard Local Search (SLS) and Variable-Depth Search (VDS) on an artificial test function. We determine features of the fitness landscape that lead to the (1+1) EA using SLS outperforming the (1+1) EA using VDS with an exponential performance gap. Moreover, we present a new local search operator, Opportunistic Local Search (OLS), that can deal with such features in the landscape and show that the (1+1) EA with OLS can efficiently optimize a discretized Rastrigin function. Stochastic fitness functions. We analyze the role of populations in stochastic optimization. We assume that the objective function is subject to noise, introducing stochastic errors in its evaluation. On classical test functions, such noise makes optimization by the simple (1+1) EA hillclimber infeasible even in exponential time. Interestingly, the use of parent and offspring populations of only logarithmic size turns the algorithm into an efficient one. The results are obtained by drift analysis. An asymptotic expansion of the expected runtime of the (1+λ) EA on ONEMAX. We consider the (1+λ) EA with mutation probability c/n, where c > 0 is a constant on ONEMAX. We give an asymptotic expansion for the expected runtime depending on both c and λ. Our results show that c = 1 is the optimal mutation rate for λ = o(lognloglogn/logloglogn) and that c only has an impact on the lower-order terms of the expected runtime, i.e. c = 1 is no longer the only optimal mutation rate. Our methods are strongly based on variable drift theorems for upper and lower bounds and a precise analysis of order statistics of the binomial distribution. To the best of our knowledge this is the first tight runtime analysis of a populationbased EA, up to lower-order terms. Furthermore, we develop helpful stochastic tools for runtime analyses. Optimal mutation rates for the (1+λ) EA on ONEMAX. We consider the (1+λ) EA with mutation probability c/n on ONEMAX, where c > 0 and λ are constant. We present an improved variable drift theorem that weakens the requirement that no large steps towards the optimum may occur in the process to a stochastic one, reducing the analysis of the expected optimization time to finding an exact expression for the drift. We formalize an exact closed-form expression for the drift and provide small error approximations that are very efficient to compute. Self-adjusting mutation rates for the (1+λ) EA on ONEMAX. We propose a new mechanism to self-adjust the mutation rate in population-based evolutionary algorithms. It consists of creating half the offspring with a higher and the rest with a lower mutation rate. The mutation rate is then adjusted, based on the success of the subpopulations. We show that the (1+λ) EA optimizes ONEMAX in an expected optimization time of O(nλ/logλ + nlogn) which has been shown to be best-possible among all λ-parallel mutation-based unbiased black-box algorithms.

[1]  Carsten Witt,et al.  (1+1) EA on Generalized Dynamic OneMax , 2015, FOGA.

[2]  Benjamin Doerr,et al.  Multiplicative drift analysis , 2010, GECCO.

[3]  Jens Jägersküpper,et al.  Combining Markov-Chain Analysis and Drift Analysis , 2011, Algorithmica.

[4]  Benjamin Doerr,et al.  k-Bit Mutation with Self-Adjusting k Outperforms Standard Bit Mutation , 2016, PPSN.

[5]  Duc-Cuong Dang,et al.  Self-adaptation of Mutation Rates in Non-elitist Populations , 2016, PPSN.

[6]  Dirk Sudholt,et al.  Hybridizing Evolutionary Algorithms with Variable-Depth Search to Overcome Local Optima , 2011, Algorithmica.

[7]  Benjamin Doerr,et al.  Optimal Parameter Settings for the (1 + λ, λ) Genetic Algorithm , 2016, GECCO.

[8]  Christine Zarges,et al.  Rigorous Runtime Analysis of Inversely Fitness Proportional Mutation Rates , 2008, PPSN.

[9]  Dirk Sudholt,et al.  Adaptive population models for offspring populations and parallel evolutionary algorithms , 2011, FOGA '11.

[10]  Duc-Cuong Dang,et al.  Level-Based Analysis of Genetic Algorithms and Other Search Processes , 2014, bioRxiv.

[11]  Benjamin Doerr,et al.  The Right Mutation Strength for Multi-Valued Decision Variables , 2016, GECCO.

[12]  Christopher R. Stephens,et al.  Rank based variation operators for genetic algorithms , 2008, GECCO '08.

[13]  Chao Qian,et al.  Running Time Analysis of the (1+1\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$1+1$$\end{document})-EA for OneMax an , 2017, Algorithmica.

[14]  Martin Dietzfelbinger,et al.  Tight Bounds for Blind Search on the Integers and the Reals , 2010, Comb. Probab. Comput..

[15]  Benjamin Doerr,et al.  Unknown solution length problems with no asymptotically optimal run time , 2017, GECCO.

[16]  Per Kristian Lehre,et al.  Concentrated Hitting Times of Randomized Search Heuristics with Variable Drift , 2014, ISAAC.

[17]  Jonathan E. Rowe,et al.  Theoretical analysis of local search strategies to optimize network communication subject to preserving the total number of links , 2009, Int. J. Intell. Comput. Cybern..

[18]  R. Kaas,et al.  Mean, Median and Mode in Binomial Distributions , 1980 .

[19]  Per Kristian Lehre,et al.  Theoretical analysis of rank-based mutation - combining exploration and exploitation , 2009, 2009 IEEE Congress on Evolutionary Computation.

[20]  Stefan Droste,et al.  A rigorous analysis of the compact genetic algorithm for linear functions , 2006, Natural Computing.

[21]  Ingo Wegener,et al.  Randomized local search, evolutionary algorithms, and the minimum spanning tree problem , 2004, Theor. Comput. Sci..

[22]  Benjamin Doerr,et al.  Runtime analysis of the (1 + (λ, λ)) genetic algorithm on random satisfiable 3-CNF formulas , 2017, GECCO.

[23]  Per Kristian Lehre,et al.  Unbiased Black-Box Complexity of Parallel Search , 2014, PPSN.

[24]  Xin Yao,et al.  On the Effectiveness of Sampling for Evolutionary Optimization in Noisy Environments , 2014, Evolutionary Computation.

[25]  Marvin Künnemann,et al.  Optimizing linear functions with the (1+λ) evolutionary algorithm - Different asymptotic runtimes for different instances , 2015, Theor. Comput. Sci..

[26]  Frank Neumann,et al.  Optimal Fixed and Adaptive Mutation Rates for the LeadingOnes Problem , 2010, PPSN.

[27]  Per Kristian Lehre,et al.  Non-uniform mutation rates for problems with unknown solution lengths , 2011, FOGA '11.

[28]  Benjamin Doerr,et al.  From black-box complexity to designing new genetic algorithms , 2015, Theor. Comput. Sci..

[29]  Pietro Simone Oliveto,et al.  How to Escape Local Optima in Black Box Optimisation: When Non-elitism Outperforms Elitism , 2017, Algorithmica.

[30]  Kenneth A. De Jong,et al.  Design and Management of Complex Technical Processes and Systems by Means of Computational Intelligence Methods on the Choice of the Offspring Population Size in Evolutionary Algorithms on the Choice of the Offspring Population Size in Evolutionary Algorithms , 2004 .

[31]  Zbigniew Michalewicz,et al.  Parameter Control in Evolutionary Algorithms , 2007, Parameter Setting in Evolutionary Algorithms.

[32]  Benjamin Doerr,et al.  Provably Optimal Self-adjusting Step Sizes for Multi-valued Decision Variables , 2016, PPSN.

[33]  Andrew M. Sutton,et al.  Robustness of Ant Colony Optimization to Noise , 2015, GECCO.

[34]  Ingo Wegener,et al.  Simulated Annealing Beats Metropolis in Combinatorial Optimization , 2005, ICALP.

[35]  Hsien-Kuei Hwang,et al.  Probabilistic Analysis of the (1+1)-Evolutionary Algorithm , 2014, Evolutionary Computation.

[36]  Benjamin Doerr,et al.  Analyzing Randomized Search Heuristics: Tools from Probability Theory , 2011, Theory of Randomized Search Heuristics.

[37]  Christine Zarges,et al.  On the utility of the population size for inversely fitness proportional mutation rates , 2009, FOGA '09.

[38]  Thomas Jansen,et al.  On the analysis of a dynamic evolutionary algorithm , 2006, J. Discrete Algorithms.

[39]  Benjamin Doerr,et al.  Fast genetic algorithms , 2017, GECCO.

[40]  Benjamin Doerr,et al.  Optimal Parameter Choices Through Self-Adjustment: Applying the 1/5-th Rule in Discrete Settings , 2015, GECCO.

[41]  Daniel Johannsen,et al.  Random combinatorial structures and randomized search heuristics , 2010 .

[42]  Carsten Witt,et al.  Tight Bounds on the Optimization Time of a Randomized Search Heuristic on Linear Functions† , 2013, Combinatorics, Probability and Computing.

[43]  Carsten Witt,et al.  The Interplay of Population Size and Mutation Probability in the (1+λ) EA on OneMax , 2017, Algorithmica.