Upper Bounds on the Running Time of the Univariate Marginal Distribution Algorithm on OneMax

The Univariate Marginal Distribution Algorithm (UMDA) is a randomized search heuristic that builds a stochastic model of the underlying optimization problem by repeatedly sampling $$\lambda $$λ solutions and adjusting the model according to the best $$\mu $$μ samples. We present a running time analysis of the UMDA on the classical OneMax benchmark function for wide ranges of the parameters $$\mu $$μ and $$\lambda $$λ. If $$\mu \ge c\log n$$μ≥clogn for some constant $$c>0$$c>0 and $$\lambda =(1+\varTheta (1))\mu $$λ=(1+Θ(1))μ, we obtain a general bound $$O(\mu n)$$O(μn) on the expected running time. This bound crucially assumes that all marginal probabilities of the algorithm are confined to the interval $$[1/n,1-1/n]$$[1/n,1-1/n]. If $$\mu \ge c' \sqrt{n}\log n$$μ≥c′nlogn for a constant $$c'>0$$c′>0 and $$\lambda =(1+\varTheta (1))\mu $$λ=(1+Θ(1))μ, the behavior of the algorithm changes and the bound on the expected running time becomes $$O(\mu \sqrt{n})$$O(μn), which typically holds even if the borders on the marginal probabilities are omitted. The results supplement the recently derived lower bound $$\varOmega (\mu \sqrt{n}+n\log n)$$Ω(μn+nlogn) by Krejca and Witt (Proceedings of FOGA 2017, ACM Press, New York, pp 65–79, 2017) and turn out to be tight for the two very different choices $$\mu =c\log n$$μ=clogn and $$\mu =c'\sqrt{n}\log n$$μ=c′nlogn. They also improve the previously best known upper bound $$O(n\log n\log \log n)$$O(nlognloglogn) by Dang and Lehre (Proceedings of GECCO ’15, ACM Press, New York, pp 513–518, 2015) that was established for $$\mu =c\log n$$μ=clogn and $$\lambda =(1+\varTheta (1))\mu $$λ=(1+Θ(1))μ.

[1]  Xin Yao,et al.  Rigorous time complexity analysis of Univariate Marginal Distribution Algorithm with margins , 2009, 2009 IEEE Congress on Evolutionary Computation.

[2]  M. T. Alodat,et al.  the Normal Distribution Function , 2008 .

[3]  Duc-Cuong Dang,et al.  Simplified Runtime Analysis of Estimation of Distribution Algorithms , 2015, GECCO.

[4]  Martin Pelikan,et al.  An introduction and survey of estimation of distribution algorithms , 2011, Swarm Evol. Comput..

[5]  Xin Yao,et al.  Analysis of Computational Time of Simple Estimation of Distribution Algorithms , 2010, IEEE Transactions on Evolutionary Computation.

[6]  Dirk Sudholt,et al.  A New Method for Lower Bounds on the Running Time of Evolutionary Algorithms , 2011, IEEE Transactions on Evolutionary Computation.

[7]  H. Mühlenbein,et al.  From Recombination of Genes to the Estimation of Distributions I. Binary Parameters , 1996, PPSN.

[8]  Pietro Simone Oliveto,et al.  Improved time complexity analysis of the Simple Genetic Algorithm , 2015, Theor. Comput. Sci..

[9]  R. Kaas,et al.  Mean, Median and Mode in Binomial Distributions , 1980 .

[10]  S. M. Samuels On the Number of Successes in Independent Trials , 1965 .

[11]  M. Habib Probabilistic methods for algorithmic discrete mathematics , 1998 .

[12]  Xin Yao,et al.  On the analysis of average time complexity of estimation of distribution algorithms , 2007, 2007 IEEE Congress on Evolutionary Computation.

[13]  William Feller,et al.  An Introduction to Probability Theory and Its Applications , 1951 .

[14]  B. Hajek Hitting-time and occupation-time bounds implied by drift analysis with applications , 1982, Advances in Applied Probability.

[15]  Carsten Witt,et al.  Tight Bounds on the Optimization Time of a Randomized Search Heuristic on Linear Functions† , 2013, Combinatorics, Probability and Computing.

[16]  Andrew Rosalsky,et al.  ON THE NUMBER OF SUCCESSES IN INDEPENDENT TRIALS , 2016 .

[17]  Carsten Witt,et al.  Lower Bounds on the Run Time of the Univariate Marginal Distribution Algorithm on OneMax , 2017, FOGA '17.

[18]  KaleJyoti,et al.  Genetic Algorithms and Evolutionary Computation , 2010 .

[19]  Roberto Cominetti,et al.  A Sharp Uniform Bound for the Distribution of Sums of Bernoulli Trials , 2008, Combinatorics, probability & computing.

[20]  Per Kristian Lehre,et al.  Concentrated Hitting Times of Randomized Search Heuristics with Variable Drift , 2014, ISAAC.

[21]  Jonathan E. Rowe,et al.  Theoretical analysis of local search strategies to optimize network communication subject to preserving the total number of links , 2009, Int. J. Intell. Comput. Cybern..

[22]  Charles M. Grinstead,et al.  Introduction to probability , 1999, Statistics for the Behavioural Sciences.

[23]  Andrew M. Sutton,et al.  The Benefit of Recombination in Noisy Evolutionary Search , 2015, GECCO.

[24]  Daniel Johannsen,et al.  Random combinatorial structures and randomized search heuristics , 2010 .

[25]  Dirk Sudholt,et al.  The choice of the offspring population size in the (1, λ) evolutionary algorithm , 2014, Theor. Comput. Sci..

[26]  Dirk Sudholt,et al.  Update Strength in EDAs and ACO: How to Avoid Genetic Drift , 2016, GECCO.

[27]  Dirk Sudholt,et al.  A few ants are enough: ACO with iteration-best update , 2010, GECCO '10.

[28]  Stefan Droste,et al.  A rigorous analysis of the compact genetic algorithm for linear functions , 2006, Natural Computing.

[29]  Michael Kolonko,et al.  Stochastic Runtime Analysis of the Cross-Entropy Algorithm , 2017, IEEE Transactions on Evolutionary Computation.

[30]  J. A. Lozano,et al.  Estimation of Distribution Algorithms: A New Tool for Evolutionary Computation , 2001 .

[31]  Per Kristian Lehre,et al.  Improved runtime bounds for the univariate marginal distribution algorithm via anti-concentration , 2017, GECCO.

[32]  Per Kristian Lehre,et al.  When is an estimation of distribution algorithm better than an evolutionary algorithm? , 2009, 2009 IEEE Congress on Evolutionary Computation.

[33]  Tobias Friedrich,et al.  EDAs cannot be Balanced and Stable , 2016, GECCO.