Exponential upper bounds for the runtime of randomized search heuristics

Abstract We argue that proven exponential upper bounds on runtimes, an established area in classic algorithms, are interesting also in heuristic search and we prove several such results. We show that any of the algorithms randomized local search, Metropolis algorithm, simulated annealing, and (1+1) evolutionary algorithm can optimize any pseudo-Boolean weakly monotonic function under a large set of noise assumptions in a runtime that is at most exponential in the problem dimension n. This drastically extends a previous such result, limited to the (1+1) EA, the LeadingOnes function, and one-bit or bit-wise prior noise with noise probability at most 1/2, and at the same time simplifies its proof. With the same general argument, among others, we also derive a sub-exponential upper bound for the runtime of the ( 1 , λ ) evolutionary algorithm on the OneMax problem when the offspring population size λ is logarithmic, but below the efficiency threshold. To show that our approach can also deal with non-trivial parent population sizes, we prove an exponential upper bound for the runtime of the mutation-based version of the simple genetic algorithm on the OneMax benchmark, matching a known exponential lower bound.

[1]  Andrew M. Sutton,et al.  Parameterized Complexity Analysis of Randomized Search Heuristics , 2019, Theory of Evolutionary Computation.

[2]  Andrew M. Sutton,et al.  When resampling to cope with noise, use median, not mean , 2019, GECCO.

[3]  Luca Maria Gambardella,et al.  A survey on metaheuristics for stochastic combinatorial optimization , 2009, Natural Computing.

[4]  Johannes Lengler,et al.  Exponential slowdown for larger populations: the (µ + 1)-EA on monotone functions , 2019, FOGA '19.

[5]  Frank Neumann,et al.  Bioinspired computation in combinatorial optimization: algorithms and their computational complexity , 2010, GECCO '12.

[6]  Andrew M. Sutton,et al.  The Compact Genetic Algorithm is Efficient Under Extreme Gaussian Noise , 2017, IEEE Transactions on Evolutionary Computation.

[7]  Benjamin Doerr,et al.  Monotonic functions in EC: anything but monotone! , 2014, GECCO.

[8]  Heinz Mühlenbein,et al.  How Genetic Algorithms Really Work: Mutation and Hillclimbing , 1992, PPSN.

[9]  Per Kristian Lehre,et al.  Negative Drift in Populations , 2010, PPSN.

[10]  Thomas Jansen,et al.  On the Optimization of Unimodal Functions with the (1 + 1) Evolutionary Algorithm , 1998, PPSN.

[11]  Benjamin Doerr,et al.  From black-box complexity to designing new genetic algorithms , 2015, Theor. Comput. Sci..

[12]  Leslie Ann Goldberg,et al.  Adaptive Drift Analysis , 2011, Algorithmica.

[13]  Carsten Witt,et al.  Tight Bounds on the Optimization Time of a Randomized Search Heuristic on Linear Functions† , 2013, Combinatorics, Probability and Computing.

[14]  Thomas Jansen,et al.  On the brittleness of evolutionary algorithms , 2007, FOGA'07.

[15]  Anne Auger,et al.  Theory of Randomized Search Heuristics: Foundations and Recent Developments , 2011, Theory of Randomized Search Heuristics.

[16]  Dirk Sudholt,et al.  A Simple Ant Colony Optimizer for Stochastic Shortest Path Problems , 2012, Algorithmica.

[17]  Hiroshi Nagamochi,et al.  Confining sets and avoiding bottleneck cases: A simple maximum independent set algorithm in degree-3 graphs , 2013, Theor. Comput. Sci..

[18]  Carsten Witt,et al.  Bioinspired Computation in Combinatorial Optimization , 2010, Bioinspired Computation in Combinatorial Optimization.

[19]  Dorian Nogneng,et al.  A new analysis method for evolutionary optimization of dynamic and noisy objective functions , 2018, GECCO.

[20]  Frank Neumann,et al.  Rigorous analyses of fitness-proportional selection for optimizing linear functions , 2008, GECCO '08.

[21]  Carsten Witt,et al.  UNIVERSITY OF DORTMUND REIHE COMPUTATIONAL INTELLIGENCE COLLABORATIVE RESEARCH CENTER 531 Design and Management of Complex Technical Processes and Systems by means of Computational Intelligence Methods Worst-Case and Average-Case Approximations by Simple Randomized Search Heuristics , 2004 .

[22]  Thomas Jansen,et al.  On the analysis of the (1+1) evolutionary algorithm , 2002, Theor. Comput. Sci..

[23]  Carsten Witt,et al.  Self-Adjusting Evolutionary Algorithms for Multimodal Optimization , 2020, Algorithmica.

[24]  Thomas Jansen,et al.  Analyzing Evolutionary Algorithms: The Computer Science Perspective , 2012 .

[25]  Dirk Sudholt,et al.  The choice of the offspring population size in the (1, λ) evolutionary algorithm , 2014, Theor. Comput. Sci..

[26]  Per Kristian Lehre,et al.  Fitness-levels for non-elitist populations , 2011, GECCO '11.

[27]  Angelika Steger,et al.  When Does Hillclimbing Fail on Monotone Functions: An entropy compression argument , 2018, ANALCO.

[28]  Andrew M. Sutton,et al.  Robustness of Ant Colony Optimization to Noise , 2015, Evolutionary Computation.

[29]  Thomas Jansen,et al.  Analyzing Evolutionary Algorithms , 2015, Natural Computing Series.

[30]  Zhi-Hua Zhou,et al.  Analyzing Evolutionary Optimization in Noisy Environments , 2013, Evolutionary Computation.

[31]  Duc-Cuong Dang,et al.  Simplified Runtime Analysis of Estimation of Distribution Algorithms , 2015, GECCO.

[32]  Benjamin Doerr,et al.  Multiplicative Up-Drift , 2019, Algorithmica.

[33]  Mojgan Pourhassan,et al.  Analysis of Evolutionary Algorithms in Dynamic and Stochastic Environments , 2018, Theory of Evolutionary Computation.

[34]  Pietro Simone Oliveto,et al.  On the analysis of the simple genetic algorithm , 2012, GECCO '12.

[35]  Fedor V. Fomin,et al.  Exact exponential algorithms , 2013, CACM.

[36]  Chao Qian,et al.  Towards a Running Time Analysis of the (1+1)-EA for OneMax and LeadingOnes Under General Bit-Wise Noise , 2018, PPSN.

[37]  Dirk Sudholt,et al.  A New Method for Lower Bounds on the Running Time of Evolutionary Algorithms , 2011, IEEE Transactions on Evolutionary Computation.

[38]  Jürgen Branke,et al.  Evolutionary optimization in uncertain environments-a survey , 2005, IEEE Transactions on Evolutionary Computation.

[39]  Karsten Weicker,et al.  Metropolis and Symmetric Functions: A Swan Song , 2009, EvoCOP.

[40]  Marc Schoenauer,et al.  Rigorous Hitting Times for Binary Mutations , 1999, Evolutionary Computation.

[41]  Stefan Droste,et al.  Design and Management of Complex Technical Processes and Systems by means of Computational Intelligence Methods Analysis of the (1+1) EA for a Noisy OneMax , 2004 .

[42]  Benjamin Doerr,et al.  Probabilistic Tools for the Analysis of Randomized Optimization Heuristics , 2018, Theory of Evolutionary Computation.

[43]  Thomas Jansen,et al.  On the analysis of a dynamic evolutionary algorithm , 2006, J. Discrete Algorithms.

[44]  Benjamin Doerr Lower Bounds for Non-Elitist Evolutionary Algorithms via Negative Multiplicative Drift , 2020, Evolutionary Computation.

[45]  Ingo Wegener,et al.  On the Optimization of Monotone Polynomials by Simple Randomized Search Heuristics , 2005, Combinatorics, Probability and Computing.

[46]  David E. Goldberg,et al.  Genetic Algorithms in Search Optimization and Machine Learning , 1988 .

[47]  Johannes Lengler,et al.  Drift Analysis , 2017, Theory of Evolutionary Computation.

[48]  Olivier Teytaud,et al.  Analysis of runtime of optimization algorithms for noisy functions over discrete codomains , 2015, Theor. Comput. Sci..

[49]  Benjamin Doerr Does comma selection help to cope with local optima? , 2020, GECCO.

[50]  Pietro Simone Oliveto,et al.  Improved time complexity analysis of the Simple Genetic Algorithm , 2015, Theor. Comput. Sci..

[51]  Mahmoud Fouz,et al.  Sharp bounds by probability-generating functions and variable drift , 2011, GECCO '11.

[52]  Dirk Sudholt,et al.  Towards a Runtime Comparison of Natural and Artificial Evolution , 2015, Algorithmica.

[53]  Chao Qian,et al.  Running Time Analysis of the (1+1\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$1+1$$\end{document})-EA for OneMax an , 2017, Algorithmica.

[54]  Dirk Sudholt,et al.  Analysing the Robustness of Evolutionary Algorithms to Noise: Refined Runtime Bounds and an Example Where Noise is Beneficial , 2018, Algorithmica.

[55]  Duc-Cuong Dang,et al.  Runtime Analysis of Non-elitist Populations: From Classical Optimisation to Partial Information , 2016, Algorithmica.

[56]  Xin Yao,et al.  Drift analysis and average time complexity of evolutionary algorithms , 2001, Artif. Intell..

[57]  Benjamin Doerr,et al.  Ants easily solve stochastic shortest path problems , 2012, GECCO '12.

[58]  Benjamin Doerr,et al.  Fast genetic algorithms , 2017, GECCO.

[59]  Benjamin Doerr Exponential Upper Bounds for the Runtime of Randomized Search Heuristics , 2020, PPSN.

[60]  Pietro Simone Oliveto,et al.  Theoretical analysis of fitness-proportional selection: landscapes and efficiency , 2009, GECCO.

[61]  Benjamin Doerr,et al.  Analyzing randomized search heuristics via stochastic domination , 2019, Theor. Comput. Sci..

[62]  Benjamin Doerr,et al.  Precise Runtime Analysis for Plateaus , 2018, PPSN.

[63]  Andrei Lissovoi,et al.  On the Time Complexity of Algorithm Selection Hyper-Heuristics for Multimodal Optimisation , 2019, AAAI.

[64]  Frank Neumann,et al.  Optimal Fixed and Adaptive Mutation Rates for the LeadingOnes Problem , 2010, PPSN.

[65]  Johannes Lengler,et al.  A General Dichotomy of Evolutionary Algorithms on Monotone Functions , 2018, IEEE Transactions on Evolutionary Computation.

[66]  Thomas Jansen,et al.  Mutation Rate Matters Even When Optimizing Monotonic Functions , 2013, Evolutionary Computation.