Domino convergence: why one should hill-climb on linear functions

In the theory community of evolutionary computation, linear pseudo-boolean functions are often regarded as easy problems since all of them can be optimized in expected time O(n log n) by simple unbiased algorithms. However, results from genetic algorithms and estimation-of-distribution algorithms indicate that these algorithms treat different linear functions differently. More precisely, an effect called "domino convergence" is described in the literature, which means that bits of large weight in the linear function are optimized earlier than bits of low weight. Hence, different linear functions may lead to rather different expected optimization times. The present paper conducts a study of domino convergence. By rigorous runtime analyses, it is shown that domino convergence is mostly a consequence of the crossover underlying genetic algorithms and EDAs. Here a performance gap of order Ω(n/log n) between different linear functions is proved. In simple mutation-only EAs the effect of domino convergence is much less pronounced, with the typical performance gap being logarithmic in the population size. The effect disappears when population size 1 is used and the algorithm is reduced to hillclimbing. Different selection mechanisms, including cut and tournament selection are investigated and their impact on domino convergence is analyzed.

[1]  Thomas Jansen,et al.  On the analysis of the (1+1) evolutionary algorithm , 2002, Theor. Comput. Sci..

[2]  Carsten Witt,et al.  Tight Bounds on the Optimization Time of a Randomized Search Heuristic on Linear Functions† , 2013, Combinatorics, Probability and Computing.

[3]  Dirk Sudholt,et al.  The choice of the offspring population size in the (1,λ) EA , 2012, GECCO '12.

[4]  Patrick J. Fitzsimmons,et al.  Converse Jensen Inequality , 2009 .

[5]  D. Goldberg,et al.  Domino convergence, drift, and the temporal-salience structure of problems , 1998, 1998 IEEE International Conference on Evolutionary Computation Proceedings. IEEE World Congress on Computational Intelligence (Cat. No.98TH8360).

[6]  David E. Goldberg,et al.  Time Complexity of genetic algorithms on exponentially scaled problems , 2000, GECCO.

[7]  David E. Goldberg,et al.  The compact genetic algorithm , 1999, IEEE Trans. Evol. Comput..

[8]  Timo Kötzing Concentration of First Hitting Times Under Additive Drift , 2015, Algorithmica.

[9]  W. Rudnick Genetic algorithms and fitness variance with an application to the automated design of artificial neural networks , 1992 .

[10]  Jens Jägersküpper,et al.  Combining Markov-Chain Analysis and Drift Analysis , 2011, Algorithmica.

[11]  Carsten Witt,et al.  Runtime Analysis of the ( μ +1) EA on Simple Pseudo-Boolean Functions , 2006 .

[12]  Andrew M. Sutton,et al.  The Benefit of Recombination in Noisy Evolutionary Search , 2015, GECCO.

[13]  Dirk Sudholt,et al.  The choice of the offspring population size in the (1, λ) evolutionary algorithm , 2014, Theor. Comput. Sci..

[14]  Annie S. Wu,et al.  A Simple Method For Detecting Domino Convergence And Identifying Salient Genes Within A Genetic Algorithm , 2002, GECCO.

[15]  Stefan Droste,et al.  A rigorous analysis of the compact genetic algorithm for linear functions , 2006, Natural Computing.

[16]  Marvin Künnemann,et al.  Optimizing linear functions with the (1+λ) evolutionary algorithm - Different asymptotic runtimes for different instances , 2015, Theor. Comput. Sci..

[17]  Benjamin Doerr,et al.  Multiplicative Drift Analysis , 2010, GECCO '10.

[18]  Dirk Sudholt,et al.  Update Strength in EDAs and ACO: How to Avoid Genetic Drift , 2016, GECCO.

[19]  Carsten Witt,et al.  Runtime Analysis of the ( + 1) EA on Simple Pseudo-Boolean Functions , 2006, Evolutionary Computation.