Exponential Upper Bounds for the Runtime of Randomized Search Heuristics

Abstract We argue that proven exponential upper bounds on runtimes, an established area in classic algorithms, are interesting also in heuristic search and we prove several such results. We show that any of the algorithms randomized local search, Metropolis algorithm, simulated annealing, and (1+1) evolutionary algorithm can optimize any pseudo-Boolean weakly monotonic function under a large set of noise assumptions in a runtime that is at most exponential in the problem dimension n. This drastically extends a previous such result, limited to the (1+1) EA, the LeadingOnes function, and one-bit or bit-wise prior noise with noise probability at most 1/2, and at the same time simplifies its proof. With the same general argument, among others, we also derive a sub-exponential upper bound for the runtime of the ( 1 , λ ) evolutionary algorithm on the OneMax problem when the offspring population size λ is logarithmic, but below the efficiency threshold. To show that our approach can also deal with non-trivial parent population sizes, we prove an exponential upper bound for the runtime of the mutation-based version of the simple genetic algorithm on the OneMax benchmark, matching a known exponential lower bound.

[1]  Avi Wigderson,et al.  An Analysis of a Simple Genetic Algorithm , 1991, ICGA.

[2]  Stefan Droste,et al.  Design and Management of Complex Technical Processes and Systems by means of Computational Intelligence Methods Analysis of the (1+1) EA for a Noisy OneMax , 2004 .

[3]  Hiroshi Nagamochi,et al.  Confining sets and avoiding bottleneck cases: A simple maximum independent set algorithm in degree-3 graphs , 2013, Theor. Comput. Sci..

[4]  Ingo Wegener,et al.  On the Optimization of Monotone Polynomials by Simple Randomized Search Heuristics , 2005, Combinatorics, Probability and Computing.

[5]  Frank Neumann,et al.  Bioinspired computation in combinatorial optimization: algorithms and their computational complexity , 2010, GECCO '12.

[6]  Carsten Witt,et al.  UNIVERSITY OF DORTMUND REIHE COMPUTATIONAL INTELLIGENCE COLLABORATIVE RESEARCH CENTER 531 Design and Management of Complex Technical Processes and Systems by means of Computational Intelligence Methods Worst-Case and Average-Case Approximations by Simple Randomized Search Heuristics , 2004 .

[7]  David E. Goldberg,et al.  Genetic Algorithms in Search Optimization and Machine Learning , 1988 .

[8]  Carsten Witt,et al.  Self-Adjusting Evolutionary Algorithms for Multimodal Optimization , 2020, Algorithmica.

[9]  Exponential slowdown for larger populations: The (μ + 1)-EA on monotone functions , 2021, Theor. Comput. Sci..

[10]  Andrew M. Sutton,et al.  The Compact Genetic Algorithm is Efficient Under Extreme Gaussian Noise , 2017, IEEE Transactions on Evolutionary Computation.

[11]  Thomas Jansen,et al.  On the analysis of a dynamic evolutionary algorithm , 2006, J. Discrete Algorithms.

[12]  Zhi-Hua Zhou,et al.  Analyzing Evolutionary Optimization in Noisy Environments , 2013, Evolutionary Computation.

[13]  Chao Qian,et al.  Running Time Analysis of the (1+1\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$1+1$$\end{document})-EA for OneMax an , 2017, Algorithmica.

[14]  Pietro Simone Oliveto,et al.  Theoretical analysis of fitness-proportional selection: landscapes and efficiency , 2009, GECCO.

[15]  Olivier Teytaud,et al.  Analysis of runtime of optimization algorithms for noisy functions over discrete codomains , 2015, Theor. Comput. Sci..

[16]  Benjamin Doerr Does Comma Selection Help to Cope with Local Optima? , 2020, GECCO.

[17]  Johannes Lengler,et al.  Exponential slowdown for larger populations: the (µ + 1)-EA on monotone functions , 2019, FOGA '19.

[18]  Frank Neumann,et al.  Optimal Fixed and Adaptive Mutation Rates for the LeadingOnes Problem , 2010, PPSN.

[19]  Duc-Cuong Dang,et al.  Runtime Analysis of Non-elitist Populations: From Classical Optimisation to Partial Information , 2016, Algorithmica.

[20]  Lower Bounds for Non-Elitist Evolutionary Algorithms Via Negative Multiplicative Drift , 2020, PPSN.

[21]  Luca Maria Gambardella,et al.  A survey on metaheuristics for stochastic combinatorial optimization , 2009, Natural Computing.

[22]  Marc Schoenauer,et al.  Rigorous Hitting Times for Binary Mutations , 1999, Evolutionary Computation.

[23]  Schloss Birlinghoven,et al.  How Genetic Algorithms Really Work I.mutation and Hillclimbing , 2022 .

[24]  Dirk Sudholt,et al.  The choice of the offspring population size in the (1, λ) evolutionary algorithm , 2014, Theor. Comput. Sci..

[25]  Fedor V. Fomin,et al.  Exact exponential algorithms , 2013, CACM.

[26]  Duc-Cuong Dang,et al.  Simplified Runtime Analysis of Estimation of Distribution Algorithms , 2015, GECCO.

[27]  Frank Neumann,et al.  Rigorous analyses of fitness-proportional selection for optimizing linear functions , 2008, GECCO '08.

[28]  Dorian Nogneng,et al.  A new analysis method for evolutionary optimization of dynamic and noisy objective functions , 2018, GECCO.

[29]  Thomas Jansen,et al.  On the brittleness of evolutionary algorithms , 2007, FOGA'07.

[30]  Andrew M. Sutton,et al.  When resampling to cope with noise, use median, not mean , 2019, GECCO.

[31]  Benjamin Doerr,et al.  Multiplicative Up-Drift , 2019, Algorithmica.

[32]  Thomas Jansen,et al.  On the Optimization of Unimodal Functions with the (1 + 1) Evolutionary Algorithm , 1998, PPSN.

[33]  Thomas Jansen,et al.  On the analysis of the (1+1) evolutionary algorithm , 2002, Theor. Comput. Sci..

[34]  Angelika Steger,et al.  When Does Hillclimbing Fail on Monotone Functions: An entropy compression argument , 2018, ANALCO.

[35]  Leslie Ann Goldberg,et al.  Adaptive Drift Analysis , 2011, Algorithmica.

[36]  Pietro Simone Oliveto,et al.  On the analysis of the simple genetic algorithm , 2012, GECCO '12.

[37]  Andrei Lissovoi,et al.  On the Time Complexity of Algorithm Selection Hyper-Heuristics for Multimodal Optimisation , 2019, AAAI.

[38]  Carsten Witt,et al.  Bioinspired Computation in Combinatorial Optimization , 2010, Bioinspired Computation in Combinatorial Optimization.

[39]  Per Kristian Lehre,et al.  Fitness-levels for non-elitist populations , 2011, GECCO '11.

[40]  Thomas Jansen,et al.  Analyzing Evolutionary Algorithms , 2015, Natural Computing Series.

[41]  Benjamin Doerr,et al.  Fast genetic algorithms , 2017, GECCO.

[42]  Mojgan Pourhassan,et al.  Analysis of Evolutionary Algorithms in Dynamic and Stochastic Environments , 2018, Theory of Evolutionary Computation.

[43]  Chao Qian,et al.  Towards a Running Time Analysis of the (1+1)-EA for OneMax and LeadingOnes Under General Bit-Wise Noise , 2018, PPSN.

[44]  Mahmoud Fouz,et al.  Sharp bounds by probability-generating functions and variable drift , 2011, GECCO '11.

[45]  Exponential Upper Bounds for the Runtime of Randomized Search Heuristics , 2020, PPSN.

[46]  Anne Auger,et al.  Theory of Randomized Search Heuristics: Foundations and Recent Developments , 2011, Theory of Randomized Search Heuristics.

[47]  Per Kristian Lehre,et al.  Negative Drift in Populations , 2010, PPSN.

[48]  Dirk Sudholt,et al.  Analysing the Robustness of Evolutionary Algorithms to Noise: Refined Runtime Bounds and an Example Where Noise is Beneficial , 2018, Algorithmica.

[49]  Dirk Sudholt,et al.  A New Method for Lower Bounds on the Running Time of Evolutionary Algorithms , 2011, IEEE Transactions on Evolutionary Computation.

[50]  Karsten Weicker,et al.  Metropolis and Symmetric Functions: A Swan Song , 2009, EvoCOP.

[51]  Dirk Sudholt,et al.  A Simple Ant Colony Optimizer for Stochastic Shortest Path Problems , 2012, Algorithmica.

[52]  Per Kristian Lehre,et al.  Drift analysis , 2012, GECCO '12.

[53]  Heinz Mühlenbein,et al.  How Genetic Algorithms Really Work: Mutation and Hillclimbing , 1992, PPSN.

[54]  Thomas Jansen,et al.  Analyzing Evolutionary Algorithms: The Computer Science Perspective , 2012 .

[55]  Benjamin Doerr,et al.  Ants easily solve stochastic shortest path problems , 2012, GECCO '12.

[56]  Dirk Sudholt,et al.  Towards a Runtime Comparison of Natural and Artificial Evolution , 2015, Algorithmica.

[57]  Andrew M. Sutton,et al.  Robustness of Ant Colony Optimization to Noise , 2015, Evolutionary Computation.

[58]  Benjamin Doerr,et al.  From black-box complexity to designing new genetic algorithms , 2015, Theor. Comput. Sci..

[59]  Benjamin Doerr,et al.  Precise Runtime Analysis for Plateaus , 2018, PPSN.

[60]  Xin Yao,et al.  Drift analysis and average time complexity of evolutionary algorithms , 2001, Artif. Intell..

[61]  Andrew M. Sutton,et al.  Parameterized Complexity Analysis of Randomized Search Heuristics , 2019, Theory of Evolutionary Computation.

[62]  Carsten Witt,et al.  Tight Bounds on the Optimization Time of a Randomized Search Heuristic on Linear Functions† , 2013, Combinatorics, Probability and Computing.

[63]  Benjamin Doerr,et al.  Monotonic functions in EC: anything but monotone! , 2014, GECCO.

[64]  Chao Qian,et al.  Running time analysis of the (1+1)-EA for onemax and leadingones under bit-wise noise , 2017, GECCO.

[65]  Johannes Lengler,et al.  A General Dichotomy of Evolutionary Algorithms on Monotone Functions , 2018, IEEE Transactions on Evolutionary Computation.

[66]  Thomas Jansen,et al.  Mutation Rate Matters Even When Optimizing Monotonic Functions , 2013, Evolutionary Computation.

[67]  Pietro Simone Oliveto,et al.  Improved time complexity analysis of the Simple Genetic Algorithm , 2015, Theor. Comput. Sci..

[68]  Benjamin Doerr,et al.  Probabilistic Tools for the Analysis of Randomized Optimization Heuristics , 2018, Theory of Evolutionary Computation.

[69]  Jürgen Branke,et al.  Evolutionary optimization in uncertain environments-a survey , 2005, IEEE Transactions on Evolutionary Computation.