Exponential Upper Bounds for the Runtime of Randomized Search Heuristics

We argue that proven exponential upper bounds on runtimes, an established area in classic algorithms, are interesting also in heuristic search and we prove several such results. We show that any of the algorithms randomized local search, Metropolis algorithm, simulated annealing, and (1+1) evolutionary algorithm can optimize any pseudo-Boolean weakly monotonic function under a large set of noise assumptions in a runtime that is at most exponential in the problem dimension~$n$. This drastically extends a previous such result, limited to the (1+1) EA, the LeadingOnes function, and one-bit or bit-wise prior noise with noise probability at most $1/2$, and at the same time simplifies its proof. With the same general argument, among others, we also derive a sub-exponential upper bound for the runtime of the $(1,\lambda)$ evolutionary algorithm on the OneMax problem when the offspring population size $\lambda$ is logarithmic, but below the efficiency threshold. To show that our approach can also deal with non-trivial parent population sizes, we prove an exponential upper bound for the runtime of the mutation-based version of the simple genetic algorithm on the OneMax benchmark, matching a known exponential lower bound.

[1]  Pietro Simone Oliveto,et al.  Theoretical analysis of fitness-proportional selection: landscapes and efficiency , 2009, GECCO.

[2]  Duc-Cuong Dang,et al.  Runtime Analysis of Non-elitist Populations: From Classical Optimisation to Partial Information , 2016, Algorithmica.

[3]  Benjamin Doerr,et al.  Analyzing randomized search heuristics via stochastic domination , 2019, Theor. Comput. Sci..

[4]  Xin Yao,et al.  Drift analysis and average time complexity of evolutionary algorithms , 2001, Artif. Intell..

[5]  Frank Neumann,et al.  Bioinspired computation in combinatorial optimization: algorithms and their computational complexity , 2010, GECCO '12.

[6]  Stefan Droste,et al.  Design and Management of Complex Technical Processes and Systems by means of Computational Intelligence Methods Analysis of the (1+1) EA for a Noisy OneMax , 2004 .

[7]  Andrew M. Sutton,et al.  Parameterized Complexity Analysis of Randomized Search Heuristics , 2019, Theory of Evolutionary Computation.

[8]  Benjamin Doerr Does comma selection help to cope with local optima? , 2020, GECCO.

[9]  Pietro Simone Oliveto,et al.  Improved time complexity analysis of the Simple Genetic Algorithm , 2015, Theor. Comput. Sci..

[10]  Benjamin Doerr,et al.  Ants easily solve stochastic shortest path problems , 2012, GECCO '12.

[11]  Johannes Lengler,et al.  Exponential slowdown for larger populations: the (µ + 1)-EA on monotone functions , 2019, FOGA '19.

[12]  Mahmoud Fouz,et al.  Sharp bounds by probability-generating functions and variable drift , 2011, GECCO '11.

[13]  Thomas Jansen,et al.  On the analysis of a dynamic evolutionary algorithm , 2006, J. Discrete Algorithms.

[14]  Ingo Wegener,et al.  On the Optimization of Monotone Polynomials by Simple Randomized Search Heuristics , 2005, Combinatorics, Probability and Computing.

[15]  Heinz Mühlenbein,et al.  How Genetic Algorithms Really Work: Mutation and Hillclimbing , 1992, PPSN.

[16]  David E. Goldberg,et al.  Genetic Algorithms in Search Optimization and Machine Learning , 1988 .

[17]  Per Kristian Lehre,et al.  Negative Drift in Populations , 2010, PPSN.

[18]  Thomas Jansen,et al.  On the Optimization of Unimodal Functions with the (1 + 1) Evolutionary Algorithm , 1998, PPSN.

[19]  Olivier Teytaud,et al.  Analysis of runtime of optimization algorithms for noisy functions over discrete codomains , 2015, Theor. Comput. Sci..

[20]  Frank Neumann,et al.  Rigorous analyses of fitness-proportional selection for optimizing linear functions , 2008, GECCO '08.

[21]  Duc-Cuong Dang,et al.  Simplified Runtime Analysis of Estimation of Distribution Algorithms , 2015, GECCO.

[22]  Dirk Sudholt,et al.  A New Method for Lower Bounds on the Running Time of Evolutionary Algorithms , 2011, IEEE Transactions on Evolutionary Computation.

[23]  Carsten Witt,et al.  UNIVERSITY OF DORTMUND REIHE COMPUTATIONAL INTELLIGENCE COLLABORATIVE RESEARCH CENTER 531 Design and Management of Complex Technical Processes and Systems by means of Computational Intelligence Methods Worst-Case and Average-Case Approximations by Simple Randomized Search Heuristics , 2004 .

[24]  Karsten Weicker,et al.  Metropolis and Symmetric Functions: A Swan Song , 2009, EvoCOP.

[25]  Benjamin Doerr,et al.  Fast genetic algorithms , 2017, GECCO.

[26]  Marc Schoenauer,et al.  Rigorous Hitting Times for Binary Mutations , 1999, Evolutionary Computation.

[27]  Chao Qian,et al.  Running time analysis of the (1+1)-EA for onemax and leadingones under bit-wise noise , 2017, GECCO.

[28]  Benjamin Doerr,et al.  Probabilistic Tools for the Analysis of Randomized Optimization Heuristics , 2018, Theory of Evolutionary Computation.

[29]  Fedor V. Fomin,et al.  Exact exponential algorithms , 2013, CACM.

[30]  Chao Qian,et al.  Towards a Running Time Analysis of the (1+1)-EA for OneMax and LeadingOnes Under General Bit-Wise Noise , 2018, PPSN.

[31]  Jürgen Branke,et al.  Evolutionary optimization in uncertain environments-a survey , 2005, IEEE Transactions on Evolutionary Computation.

[32]  Schloss Birlinghoven,et al.  How Genetic Algorithms Really Work I.mutation and Hillclimbing , 2022 .

[33]  Dirk Sudholt,et al.  The choice of the offspring population size in the (1, λ) evolutionary algorithm , 2014, Theor. Comput. Sci..

[34]  Zhi-Hua Zhou,et al.  Analyzing Evolutionary Optimization in Noisy Environments , 2013, Evolutionary Computation.

[35]  Benjamin Doerr,et al.  From black-box complexity to designing new genetic algorithms , 2015, Theor. Comput. Sci..

[36]  Thomas Jansen,et al.  On the analysis of the (1+1) evolutionary algorithm , 2002, Theor. Comput. Sci..

[37]  Carsten Witt,et al.  Self-Adjusting Evolutionary Algorithms for Multimodal Optimization , 2020, Algorithmica.

[38]  Mojgan Pourhassan,et al.  Analysis of Evolutionary Algorithms in Dynamic and Stochastic Environments , 2018, Theory of Evolutionary Computation.

[39]  Pietro Simone Oliveto,et al.  On the analysis of the simple genetic algorithm , 2012, GECCO '12.

[40]  Benjamin Doerr Lower Bounds for Non-Elitist Evolutionary Algorithms via Negative Multiplicative Drift , 2020, Evolutionary Computation.

[41]  Angelika Steger,et al.  When Does Hillclimbing Fail on Monotone Functions: An entropy compression argument , 2018, ANALCO.

[42]  Andrew M. Sutton,et al.  The Compact Genetic Algorithm is Efficient Under Extreme Gaussian Noise , 2017, IEEE Transactions on Evolutionary Computation.

[43]  Dirk Sudholt,et al.  Towards a Runtime Comparison of Natural and Artificial Evolution , 2015, Algorithmica.

[44]  Chao Qian,et al.  Running Time Analysis of the (1+1\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$1+1$$\end{document})-EA for OneMax an , 2017, Algorithmica.

[45]  Dirk Sudholt,et al.  Analysing the Robustness of Evolutionary Algorithms to Noise: Refined Runtime Bounds and an Example Where Noise is Beneficial , 2018, Algorithmica.

[46]  Thomas Jansen,et al.  Mutation Rate Matters Even When Optimizing Monotonic Functions , 2013, Evolutionary Computation.

[47]  Andrei Lissovoi,et al.  On the Time Complexity of Algorithm Selection Hyper-Heuristics for Multimodal Optimisation , 2019, AAAI.

[48]  Frank Neumann,et al.  Optimal Fixed and Adaptive Mutation Rates for the LeadingOnes Problem , 2010, PPSN.

[49]  Johannes Lengler,et al.  A General Dichotomy of Evolutionary Algorithms on Monotone Functions , 2018, IEEE Transactions on Evolutionary Computation.

[50]  Benjamin Doerr,et al.  Monotonic functions in EC: anything but monotone! , 2014, GECCO.

[51]  Per Kristian Lehre,et al.  Fitness-levels for non-elitist populations , 2011, GECCO '11.

[52]  Andrew M. Sutton,et al.  Robustness of Ant Colony Optimization to Noise , 2015, Evolutionary Computation.

[53]  Benjamin Doerr,et al.  Precise Runtime Analysis for Plateaus , 2018, PPSN.

[54]  Hiroshi Nagamochi,et al.  Confining sets and avoiding bottleneck cases: A simple maximum independent set algorithm in degree-3 graphs , 2013, Theor. Comput. Sci..

[55]  Dorian Nogneng,et al.  A new analysis method for evolutionary optimization of dynamic and noisy objective functions , 2018, GECCO.

[56]  Andrew M. Sutton,et al.  When resampling to cope with noise, use median, not mean , 2019, GECCO.

[57]  Luca Maria Gambardella,et al.  A survey on metaheuristics for stochastic combinatorial optimization , 2009, Natural Computing.

[58]  Leslie Ann Goldberg,et al.  Adaptive Drift Analysis , 2011, Algorithmica.

[59]  Thomas Jansen,et al.  On the brittleness of evolutionary algorithms , 2007, FOGA'07.

[60]  Per Kristian Lehre,et al.  Drift analysis , 2012, GECCO '12.

[61]  Thomas Jansen,et al.  Analyzing Evolutionary Algorithms: The Computer Science Perspective , 2012 .

[62]  Carsten Witt,et al.  Tight Bounds on the Optimization Time of a Randomized Search Heuristic on Linear Functions† , 2013, Combinatorics, Probability and Computing.

[63]  Dirk Sudholt,et al.  A Simple Ant Colony Optimizer for Stochastic Shortest Path Problems , 2012, Algorithmica.