Enhancing the Performance of Maximum-Likelihood Gaussian EDAs Using Anticipated Mean Shift

Many Estimation---of---Distribution Algorithms use maximum-likelihood (ML) estimates. For discrete variables this has met with great success. For continuous variables the use of ML estimates for the normal distribution does not directly lead to successful optimization in most landscapes. It was previously found that an important reason for this is the premature shrinking of the variance at an exponential rate. Remedies were subsequently successfully formulated (i.e. Adaptive Variance Scaling (AVS) and Standard---Deviation Ratio triggering (SDR)). Here we focus on a second source of inefficiency that is not removed by existing remedies. We then provide a simple, but effective technique called Anticipated Mean Shift (AMS) that removes this inefficiency.

[1]  David B. Fogel,et al.  A Note on the Empirical Evaluation of Intermediate Recombination , 1995, Evolutionary Computation.

[2]  Michèle Sebag,et al.  Extending Population-Based Incremental Learning to Continuous Search Spaces , 1998, PPSN.

[3]  Thomas Bäck,et al.  Parallel Problem Solving from Nature — PPSN V , 1998, Lecture Notes in Computer Science.

[4]  Michael I. Jordan Graphical Models , 2003 .

[5]  Pedro Larrañaga,et al.  Optimization in Continuous Domains by Learning and Simulation of Gaussian Networks , 2000 .

[6]  Dirk Thierens,et al.  Expanding from Discrete to Continuous Estimation of Distribution Algorithms: The IDEA , 2000, PPSN.

[7]  Xin Yao,et al.  Parallel Problem Solving from Nature PPSN VI , 2000, Lecture Notes in Computer Science.

[8]  Kalyanmoy Deb,et al.  On self-adaptive features in real-parameter evolutionary algorithms , 2001, IEEE Trans. Evol. Comput..

[9]  Pedro Larrañaga,et al.  Mathematical modelling of UMDAc algorithm with tournament selection. Behaviour on linear and quadratic functions , 2002, Int. J. Approx. Reason..

[10]  Petros Koumoutsakos,et al.  Reducing the Time Complexity of the Derandomized Evolution Strategy with Covariance Matrix Adaptation (CMA-ES) , 2003, Evolutionary Computation.

[11]  Petros Koumoutsakos,et al.  A Mixed Bayesian Optimization Algorithm with Variance Adaptation , 2004, PPSN.

[12]  U. Aickelin,et al.  Parallel Problem Solving from Nature - PPSN VIII , 2004, Lecture Notes in Computer Science.

[13]  Franz Rothlauf,et al.  Behaviour of UMDA/sub c/ with truncation selection on monotonous functions , 2005, 2005 IEEE Congress on Evolutionary Computation.

[14]  Marcus Gallagher,et al.  Population-Based Continuous Optimization, Probabilistic Modelling and Mean Shift , 2005, Evolutionary Computation.

[15]  Heinz Mühlenbein,et al.  The Estimation of Distributions and the Minimum Relative Entropy Principle , 2005, Evol. Comput..

[16]  Franz Rothlauf,et al.  The correlation-triggered adaptive variance scaling IDEA , 2006, GECCO.

[17]  Martin Pelikan,et al.  Scalable Optimization via Probabilistic Modeling , 2006, Studies in Computational Intelligence.

[18]  Martin Pelikan,et al.  Scalable Optimization via Probabilistic Modeling: From Algorithms to Applications (Studies in Computational Intelligence) , 2006 .

[19]  J. A. Lozano,et al.  Towards a New Evolutionary Computation: Advances on Estimation of Distribution Algorithms (Studies in Fuzziness and Soft Computing) , 2006 .

[20]  Pedro Larrañaga,et al.  Towards a New Evolutionary Computation - Advances in the Estimation of Distribution Algorithms , 2006, Towards a New Evolutionary Computation.

[21]  P. Bosman,et al.  Adapted Maximum-Likelihood Gaussian Models for Numerical Optimization with Continuous EDAs , 2007 .

[22]  Franz Rothlauf,et al.  SDR: a better trigger for adaptive variance scaling in normal EDAs , 2007, GECCO '07.

[23]  Hua Xu,et al.  Cross entropy and adaptive variance scaling in continuous EDA , 2007, GECCO '07.