The Complex Parameter Landscape of the Compact Genetic Algorithm

The compact Genetic Algorithm (cGA) evolves a probability distribution favoring optimal solutions in the underlying search space by repeatedly sampling from the distribution and updating it according to promising samples. We study the intricate dynamics of the cGA on the test function OneMax, and how its performance depends on the hypothetical population size K, which determines how quickly decisions about promising bit values are fixated in the probabilistic model. It is known that the cGA and the Univariate Marginal Distribution Algorithm (UMDA), a related algorithm whose population size is called  $$\lambda$$ , run in expected time $$O(n \log n)$$ when the population size is just large enough ( $$K = \varTheta (\sqrt{n}\log n)$$ and $$\lambda = \varTheta (\sqrt{n}\log n)$$ , respectively) to avoid wrong decisions being fixated. The UMDA also shows the same performance in a very different regime ( $$\lambda =\varTheta (\log n)$$ , equivalent to $$K = \varTheta (\log n)$$ in the cGA) with much smaller population size, but for very different reasons: many wrong decisions are fixated initially, but then reverted efficiently. If the population size is even smaller ( $$o(\log n)$$ ), the time is exponential. We show that population sizes in between the two optimal regimes are worse as they yield larger runtimes: we prove a lower bound of $$\varOmega (K^{1/3}n + n \log n)$$ for the cGA on OneMax for $$K = O(\sqrt{n}/\log ^2 n)$$ . For $$K = \varOmega (\log ^3 n)$$ the runtime increases with growing K before dropping again to $$O(K\sqrt{n} + n \log n)$$ for $$K = \varOmega (\sqrt{n} \log n)$$ . This suggests that the expected runtime for the cGA is a bimodal function in K with two very different optimal regions and worse performance in between.

[1]  Xin Yao,et al.  A study of drift analysis for estimating computation time of evolutionary algorithms , 2004, Natural Computing.

[2]  Pietro Simone Oliveto,et al.  Simplified Drift Analysis for Proving Lower Bounds in Evolutionary Computation , 2008, Algorithmica.

[3]  Andrew M. Sutton,et al.  The Compact Genetic Algorithm is Efficient Under Extreme Gaussian Noise , 2017, IEEE Transactions on Evolutionary Computation.

[4]  Stefan Droste,et al.  A rigorous analysis of the compact genetic algorithm for linear functions , 2006, Natural Computing.

[5]  David E. Goldberg,et al.  The compact genetic algorithm , 1999, IEEE Trans. Evol. Comput..

[6]  Roberto Cominetti,et al.  A Sharp Uniform Bound for the Distribution of Sums of Bernoulli Trials , 2008, Combinatorics, probability & computing.

[7]  David E. Goldberg,et al.  The Gambler's Ruin Problem, Genetic Algorithms, and the Sizing of Populations , 1999, Evolutionary Computation.

[8]  Dirk Sudholt,et al.  The choice of the offspring population size in the (1, λ) evolutionary algorithm , 2014, Theor. Comput. Sci..

[9]  Pietro Simone Oliveto,et al.  Erratum: Simplified Drift Analysis for Proving Lower Bounds in Evolutionary Computation , 2008, PPSN.

[10]  Carsten Witt,et al.  Lower Bounds on the Run Time of the Univariate Marginal Distribution Algorithm on OneMax , 2017, FOGA '17.

[11]  Per Kristian Lehre,et al.  Improved runtime bounds for the univariate marginal distribution algorithm via anti-concentration , 2017, GECCO.

[12]  Timo Kötzing Concentration of First Hitting Times Under Additive Drift , 2015, Algorithmica.

[13]  Alessandro Panconesi,et al.  Concentration of Measure for the Analysis of Randomized Algorithms , 2009 .

[14]  William Feller,et al.  An Introduction to Probability Theory and Its Applications , 1967 .

[15]  Dirk Sudholt,et al.  Medium step sizes are harmful for the compact genetic algorithm , 2018, GECCO.

[16]  Duc-Cuong Dang,et al.  Level-Based Analysis of the Univariate Marginal Distribution Algorithm , 2018, Algorithmica.

[17]  Carsten Witt,et al.  Upper Bounds on the Running Time of the Univariate Marginal Distribution Algorithm on OneMax , 2018, Algorithmica.

[18]  Dirk Sudholt,et al.  On the Choice of the Update Strength in Estimation-of-Distribution Algorithms and Ant Colony Optimization , 2018, Algorithmica.

[19]  Dirk Sudholt,et al.  A few ants are enough: ACO with iteration-best update , 2010, GECCO '10.

[20]  Carsten Witt,et al.  Theory of estimation-of-distribution algorithms , 2018, GECCO.

[21]  Heinz Mühlenbein,et al.  Predictive Models for the Breeder Genetic Algorithm I. Continuous Parameter Optimization , 1993, Evolutionary Computation.

[22]  Dirk Sudholt,et al.  Update Strength in EDAs and ACO: How to Avoid Genetic Drift , 2016, GECCO.

[23]  Devdatt P. Dubhashi,et al.  Concentration of Measure for the Analysis of Randomized Algorithms: Contents , 2009 .