Using Previous Models to Bias Structural Learning in the Hierarchical BOA

Estimation of distribution algorithms (EDAs) are stochastic optimization techniques that explore the space of potential solutions by building and sampling explicit probabilistic models of promising candidate solutions. While the primary goal of applying EDAs is to discover the global optimum or at least its accurate approximation, besides this, any EDA provides us with a sequence of probabilistic models, which in most cases hold a great deal of information about the problem. Although using problem-specific knowledge has been shown to significantly improve performance of EDAs and other evolutionary algorithms, this readily available source of problem-specific information has been practically ignored by the EDA community. This paper takes the first step toward the use of probabilistic models obtained by EDAs to speed up the solution of similar problems in the future. More specifically, we propose two approaches to biasing model building in the hierarchical Bayesian optimization algorithm (hBOA) based on knowledge automatically learned from previous hBOA runs on similar problems. We show that the proposed methods lead to substantial speedups and argue that the methods should work well in other applications that require solving a large number of problems with similar structure.

[1]  Pedro Larrañaga,et al.  Estimation of Distribution Algorithms , 2002, Genetic Algorithms and Evolutionary Computation.

[2]  H. Mühlenbein,et al.  From Recombination of Genes to the Estimation of Distributions I. Binary Parameters , 1996, PPSN.

[3]  Martin Pelikan,et al.  Scalable Optimization via Probabilistic Modeling: From Algorithms to Applications (Studies in Computational Intelligence) , 2006 .

[4]  Elena Marchiori,et al.  Evolutionary Algorithms for the Satisfiability Problem , 2002, Evolutionary Computation.

[5]  Alexander Mendiburu,et al.  Implementation and Performance Evaluation of a Parallelization of Estimation of Bayesian Network Algorithms , 2006, Parallel Process. Lett..

[6]  David E. Goldberg,et al.  A hierarchy machine: Learning to optimize from nature and humans , 2003, Complex..

[7]  P. Bosman,et al.  Continuous iterated density estimation evolutionary algorithms within the IDEA framework , 2000 .

[8]  Thomas Stützle,et al.  SATLIB: An Online Resource for Research on SAT , 2000 .

[9]  Jack Mostow,et al.  Direct Transfer of Learned Information Among Neural Networks , 1991, AAAI.

[10]  Shumeet Baluja,et al.  A Method for Integrating Genetic Search Based Function Optimization and Competitive Learning , 1994 .

[11]  Franz Rothlauf,et al.  Evaluation-Relaxation Schemes for Genetic and Evolutionary Algorithms , 2004 .

[12]  David E. Goldberg,et al.  iBOA: the incremental bayesian optimization algorithm , 2008, GECCO '08.

[13]  Martin Pelikan,et al.  Analyzing probabilistic models in hierarchical BOA on traps and spin glasses , 2007, GECCO '07.

[14]  D. Goldberg,et al.  Escaping hierarchical traps with competent genetic algorithms , 2001 .

[15]  Sebastian Thrun,et al.  Is Learning The n-th Thing Any Easier Than Learning The First? , 1995, NIPS.

[16]  Kumara Sastry,et al.  Efficient Atomic Cluster Optimization Using A Hybrid Extended Compact Genetic Algorithm With Seeded , 2001 .

[17]  Judea Pearl,et al.  Probabilistic reasoning in intelligent systems - networks of plausible inference , 1991, Morgan Kaufmann series in representation and reasoning.

[18]  Rich Caruana,et al.  Multitask Learning , 1997, Machine-mediated learning.

[19]  David E. Goldberg,et al.  Designing Competent Mutation Operators Via Probabilistic Model Building of Neighborhoods , 2004, GECCO.

[20]  M. Troyer,et al.  Performance limitations of flat-histogram methods. , 2003, Physical review letters.

[21]  Martin Pelikan,et al.  Design of Parallel Estimation of Distribution Algorithms , 2006, Scalable Optimization via Probabilistic Modeling.

[22]  Habiba Drias,et al.  A Performance Comparison of Evolutionary Meta-heuristics and Solving MAX-SAT Problems , 2004, International Conference on Computational Intelligence.

[23]  David E. Goldberg,et al.  Hierarchical BOA Solves Ising Spin Glasses and MAXSAT , 2003, GECCO.

[24]  Heinz Mühlenbein,et al.  Evolutionary optimization and the estimation of search distributions with applications to graph bipartitioning , 2002, Int. J. Approx. Reason..

[25]  David E. Goldberg,et al.  The Design of Innovation: Lessons from and for Competent Genetic Algorithms , 2002 .

[26]  David E. Goldberg,et al.  A Survey of Optimization by Building and Using Probabilistic Models , 2002, Comput. Optim. Appl..

[27]  David H. Ackley,et al.  An empirical study of bit vector function optimization , 1987 .

[28]  A. Young,et al.  Spin glasses and random fields , 1997 .

[29]  Kalyanmoy Deb,et al.  Analyzing Deception in Trap Functions , 1992, FOGA.

[30]  Martin Pelikan,et al.  Fitness Inheritance in the Bayesian Optimization Algorithm , 2004, GECCO.

[31]  Ronald A. Howard,et al.  Influence Diagrams , 2005, Decis. Anal..

[32]  A. Hartmann Cluster – Exact Approximation of Spin Glass , 2008 .

[33]  D. Goldberg,et al.  Don't evaluate, inherit , 2001 .

[34]  David E. Goldberg,et al.  Effects of a deterministic hill climber on hBOA , 2009, GECCO.

[35]  Josef Schwarz,et al.  The Parallel Bayesian Optimization Algorithm , 2000 .

[36]  Gregory F. Cooper,et al.  A Bayesian Method for the Induction of Probabilistic Networks from Data , 1992 .

[37]  J. A. Lozano,et al.  Estimation of Distribution Algorithms: A New Tool for Evolutionary Computation , 2001 .

[38]  John A. W. McCall,et al.  Solving the MAXSAT problem using a multivariate EDA based on Markov networks , 2007, GECCO '07.

[39]  Toby Walsh,et al.  Morphing: Combining Structure and Randomness , 1999, AAAI/IAAI.

[40]  Erick Cantú-Paz,et al.  Efficient and Accurate Parallel Genetic Algorithms , 2000, Genetic Algorithms and Evolutionary Computation.

[41]  K. Binder,et al.  Spin glasses: Experimental facts, theoretical concepts, and open questions , 1986 .

[42]  David E. Goldberg,et al.  Sporadic model building for efficiency enhancement of hierarchical BOA , 2006, GECCO '06.

[43]  Lorien Y. Pratt,et al.  Non-literal Transfer Among Neural Network Learners , 1993 .

[44]  David E. Goldberg,et al.  Efficiency Enhancement of Estimation of Distribution Algorithms , 2006, Scalable Optimization via Probabilistic Modeling.

[45]  L. Darrell Whitley,et al.  Genetic Algorithm Behavior in the MAXSAT Domain , 1998, PPSN.

[46]  Franz Rothlauf,et al.  Representations for genetic and evolutionary algorithms , 2002, Studies in Fuzziness and Soft Computing.

[47]  David Maxwell Chickering,et al.  A Bayesian Approach to Learning Bayesian Networks with Local Structure , 1997, UAI.

[48]  Jiri Ocenasek,et al.  Parallel Estimation of Distribution Algorithms , 2010 .

[49]  Nir Friedman,et al.  Learning Bayesian Networks with Local Structure , 1996, UAI.

[50]  Robert E. Smith,et al.  Fitness inheritance in genetic algorithms , 1995, SAC '95.

[51]  Martin Pelikan,et al.  Hierarchical Bayesian optimization algorithm: toward a new generation of evolutionary algorithms , 2010, SICE 2003 Annual Conference (IEEE Cat. No.03TH8734).

[52]  Tony Jebara,et al.  Multi-task feature and kernel selection for SVMs , 2004, ICML.

[53]  M. Mézard,et al.  Spin Glass Theory and Beyond , 1987 .

[54]  F. Guerra Spin Glasses , 2005, cond-mat/0507581.

[55]  Rich Caruana,et al.  Inductive Transfer for Bayesian Network Structure Learning , 2007, ICML Unsupervised and Transfer Learning.

[56]  Sebastian Thrun,et al.  Learning One More Thing , 1994, IJCAI.

[57]  Martin Pelikan,et al.  Scalable Optimization via Probabilistic Modeling , 2006, Studies in Computational Intelligence.

[58]  David E. Goldberg Using Time Efficiently: Genetic-Evolutionary Algorithms and the Continuation Problem , 1999, GECCO.

[59]  Martin Pelikan,et al.  Searching for Ground States of Ising Spin Glasses with Hierarchical BOA and Cluster Exact Approximation , 2006, Scalable Optimization via Probabilistic Modeling.

[60]  Georges R. Harik,et al.  Finding Multimodal Solutions Using Restricted Tournament Selection , 1995, ICGA.

[61]  David E. Goldberg,et al.  Scalability of the Bayesian optimization algorithm , 2002, Int. J. Approx. Reason..

[62]  Neil D. Lawrence,et al.  Learning to learn with the informative vector machine , 2004, ICML.

[63]  Sebastian Thrun,et al.  Lifelong Learning: A Case Study. , 1995 .

[64]  Shumeet Baluja,et al.  Incorporating a priori Knowledge in Probabilistic-Model Based Optimization , 2006, Scalable Optimization via Probabilistic Modeling.