Probability Collectives for Optimization of Computer Simulations

A significant body of research under the rubric of blackbox optimization addresses the problem of optimization in a design space where the designs are evaluated by a computer simulation. For many such simulations, conventional local optimization methods prove inadequate, owing to the presence of local minima, local non-smoothness, and so on. A more effective approach would learn the global characteristics of the design space, and sample from disparate regions as it progresses. Several probabilistic approaches to this problem are based on the following idea: during the course of any search process, our partial knowledge of the design space grows as we evaluate a growing number of designs, and this knowledge can be elegantly expressed in terms of probability distributions. This is usually done using surrogate models, or data fits. We would like to use this partial knowledge to lead us to promising new designs, and this involves searching the data fit for points where the uncertainty of the fit, in conjunction with the fit itself, indicates a good chance of finding a improved design. This is the basis for the well-known Efficient Global Optimization (EGO) algorithm. This task can be regarded as an auxiliary optimization problem, but one that has some undesirable characteristics, including large flat regions and multiple local optima. Previous solution approaches have included branch and bound (the EGO algorithm), and local optimization with multi-start. In this paper, we present a new approach, itself based on probability distributions. Probability Collectives (PC) is an optimization framework in which a given optimization problem is transformed to one over probability distributions. This transformation often enables us to overcome problems such as local non-smoothness and multiple local optima. One particular PC approach is closely related to function smoothing and filtering. Our formulation is a type of Monte Carlo Optimization, often encountered in machine learning algorithms. We briefly introduce this approach, and apply it to the auxiliary optimization problem described above. Finally, we compare this technique with other algorithms, including a heuristic sampling algorithm, and a genetic algorithm.

[1]  H. Zimmermann Towards global optimization 2: L.C.W. DIXON and G.P. SZEGÖ (eds.) North-Holland, Amsterdam, 1978, viii + 364 pages, US $ 44.50, Dfl. 100,-. , 1979 .

[2]  Ilan Kroo,et al.  Distributed optimization and flight control using collectives , 2005 .

[3]  Carl E. Rasmussen,et al.  Gaussian processes for machine learning , 2005, Adaptive computation and machine learning.

[4]  Donald R. Jones,et al.  A Taxonomy of Global Optimization Methods Based on Response Surfaces , 2001, J. Glob. Optim..

[5]  Donald R. Jones,et al.  Efficient Global Optimization of Expensive Black-Box Functions , 1998, J. Glob. Optim..

[6]  David H. Wolpert,et al.  Flight Control with Distributed Efiectors , 2005 .

[7]  Bernardetta Addis,et al.  Local optima smoothing for global optimization , 2005, Optim. Methods Softw..

[8]  Thomas J. Santner,et al.  The Design and Analysis of Computer Experiments , 2003, Springer Series in Statistics.

[9]  David H. Wolpert,et al.  Product distribution theory for control of multi-agent systems , 2004, Proceedings of the Third International Joint Conference on Autonomous Agents and Multiagent Systems, 2004. AAMAS 2004..

[10]  David H. Wolpert,et al.  Parametric Learning and Monte Carlo Optimization , 2007, ArXiv.

[11]  Douglas C. Montgomery,et al.  Response Surface Methodology: Process and Product Optimization Using Designed Experiments , 1995 .

[12]  N. M. Alexandrov,et al.  A trust-region framework for managing the use of approximation models in optimization , 1997 .

[13]  David H. Wolpert,et al.  Advances in Distributed Optimization Using Probability Collectives , 2006, Adv. Complex Syst..

[14]  A. J. Booker,et al.  A rigorous framework for optimization of expensive functions by surrogates , 1998 .

[15]  Melanie Mitchell,et al.  An introduction to genetic algorithms , 1996 .

[16]  David H. Wolpert,et al.  Information Theory - The Bridge Connecting Bounded Rational Game Theory and Statistical Physics , 2004, ArXiv.

[17]  David H. Wolpert,et al.  Distributed control by Lagrangian steepest descent , 2004, 2004 43rd IEEE Conference on Decision and Control (CDC) (IEEE Cat. No.04CH37601).

[18]  Daniel M. Dunlavy,et al.  Formulations for Surrogate-Based Optimization with Data Fit, Multifidelity, and Reduced-Order Models , 2006 .