On the Choice of the Update Strength in Estimation-of-Distribution Algorithms and Ant Colony Optimization

Probabilistic model-building Genetic Algorithms (PMBGAs) are a class of metaheuristics that evolve probability distributions favoring optimal solutions in the underlying search space by repeatedly sampling from the distribution and updating it according to promising samples. We provide a rigorous runtime analysis concerning the update strength, a vital parameter in PMBGAs such as the step size 1 / K in the so-called compact Genetic Algorithm (cGA) and the evaporation factor $$\rho $$ρ in ant colony optimizers (ACO). While a large update strength is desirable for exploitation, there is a general trade-off: too strong updates can lead to unstable behavior and possibly poor performance. We demonstrate this trade-off for the cGA and a simple ACO algorithm on the well-known OneMax function. More precisely, we obtain lower bounds on the expected runtime of $${\varOmega }(K\sqrt{n} + n \log n)$$Ω(Kn+nlogn) and $${\varOmega }(\sqrt{n}/\rho + n \log n)$$Ω(n/ρ+nlogn), respectively, suggesting that the update strength should be limited to $$1/K, \rho = O(1/(\sqrt{n} \log n))$$1/K,ρ=O(1/(nlogn)). In fact, choosing $$1/K, \rho \sim 1/(\sqrt{n}\log n)$$1/K,ρ∼1/(nlogn) both algorithms efficiently optimize OneMax in expected time $${\varTheta }(n \log n)$$Θ(nlogn). Our analyses provide new insights into the stochastic behavior of PMBGAs and propose new guidelines for setting the update strength in global optimization.

[1]  Duc-Cuong Dang,et al.  Simplified Runtime Analysis of Estimation of Distribution Algorithms , 2015, GECCO.

[2]  Andrew M. Sutton,et al.  The Compact Genetic Algorithm is Efficient Under Extreme Gaussian Noise , 2017, IEEE Transactions on Evolutionary Computation.

[3]  Thomas Stützle,et al.  MAX-MIN Ant System , 2000, Future Gener. Comput. Syst..

[4]  Per Kristian Lehre,et al.  General Drift Analysis with Tail Bounds , 2013, ArXiv.

[5]  Benjamin Doerr,et al.  Drift analysis and linear functions revisited , 2010, IEEE Congress on Evolutionary Computation.

[6]  Martin Pelikan,et al.  An introduction and survey of estimation of distribution algorithms , 2011, Swarm Evol. Comput..

[7]  Carola Doerr,et al.  OneMax in Black-Box Models with Several Restrictions , 2015, Algorithmica.

[8]  Carsten Witt,et al.  Tight Bounds on the Optimization Time of a Randomized Search Heuristic on Linear Functions† , 2013, Combinatorics, Probability and Computing.

[9]  Per Kristian Lehre,et al.  Improved runtime bounds for the univariate marginal distribution algorithm via anti-concentration , 2017, GECCO.

[10]  Dirk Sudholt,et al.  A few ants are enough: ACO with iteration-best update , 2010, GECCO '10.

[11]  Andrew M. Sutton,et al.  The Benefit of Recombination in Noisy Evolutionary Search , 2015, GECCO.

[12]  Dirk Sudholt,et al.  The choice of the offspring population size in the (1, λ) evolutionary algorithm , 2014, Theor. Comput. Sci..

[13]  Pietro Simone Oliveto,et al.  Improved time complexity analysis of the Simple Genetic Algorithm , 2015, Theor. Comput. Sci..

[14]  Per Kristian Lehre,et al.  Concentrated Hitting Times of Randomized Search Heuristics with Variable Drift , 2014, ISAAC.

[15]  David E. Goldberg,et al.  The compact genetic algorithm , 1999, IEEE Trans. Evol. Comput..

[16]  Dirk Sudholt,et al.  Analysis of different MMAS ACO algorithms on unimodal functions and plateaus , 2009, Swarm Intelligence.

[17]  Roberto Cominetti,et al.  A Sharp Uniform Bound for the Distribution of Sums of Bernoulli Trials , 2008, Combinatorics, probability & computing.

[18]  Stefan Droste,et al.  A rigorous analysis of the compact genetic algorithm for linear functions , 2006, Natural Computing.

[19]  Russ Bubley,et al.  Randomized algorithms , 1995, CSUR.

[20]  Thomas Jansen,et al.  UNIVERSITY OF DORTMUND REIHE COMPUTATIONAL INTELLIGENCE COLLABORATIVE RESEARCH CENTER 531 Design and Management of Complex Technical Processes and Systems by means of Computational Intelligence Methods Upper and Lower Bounds for Randomized Search Heuristics in Black-Box Optimization , 2004 .

[21]  I. Olkin,et al.  Inequalities: Theory of Majorization and Its Applications , 1980 .

[22]  Carsten Witt,et al.  Lower Bounds on the Run Time of the Univariate Marginal Distribution Algorithm on OneMax , 2017, FOGA '17.

[23]  Tobias Friedrich,et al.  EDAs cannot be Balanced and Stable , 2016, GECCO.

[24]  Carsten Witt,et al.  Upper Bounds on the Running Time of the Univariate Marginal Distribution Algorithm on OneMax , 2018, Algorithmica.

[25]  Dirk Sudholt,et al.  Update Strength in EDAs and ACO: How to Avoid Genetic Drift , 2016, GECCO.

[26]  Per Kristian Lehre,et al.  When is an estimation of distribution algorithm better than an evolutionary algorithm? , 2009, 2009 IEEE Congress on Evolutionary Computation.

[27]  J. Darroch On the Distribution of the Number of Successes in Independent Trials , 1964 .

[28]  Dirk Sudholt,et al.  A New Method for Lower Bounds on the Running Time of Evolutionary Algorithms , 2011, IEEE Transactions on Evolutionary Computation.

[29]  Benjamin Doerr,et al.  Analyzing Randomized Search Heuristics: Tools from Probability Theory , 2011, Theory of Randomized Search Heuristics.

[30]  Daniel Johannsen,et al.  Random combinatorial structures and randomized search heuristics , 2010 .

[31]  William Feller,et al.  An Introduction to Probability Theory and Its Applications , 1967 .