Batch Bayesian Optimization via Simulation Matching

Bayesian optimization methods are often used to optimize unknown functions that are costly to evaluate. Typically, these methods sequentially select inputs to be evaluated one at a time based on a posterior over the unknown function that is updated after each evaluation. In many applications, however, it is desirable to perform multiple evaluations in parallel, which requires selecting batches of multiple inputs to evaluate at once. In this paper, we propose a novel approach to batch Bayesian optimization, providing a policy for selecting batches of inputs with the goal of optimizing the function as efficiently as possible. The key idea is to exploit the availability of high-quality and efficient sequential policies, by using Monte-Carlo simulation to select input batches that closely match their expected behavior. Our experimental results on six benchmarks show that the proposed approach significantly outperforms two baselines and can lead to large advantages over a top sequential approach in terms of performance per unit time.

[1]  Hrishikesh D. Vinod Mathematica Integer Programming and the Theory of Grouping , 1969 .

[2]  M. L. Fisher,et al.  An analysis of approximations for maximizing submodular set functions—I , 1978, Math. Program..

[3]  S. P. Lloyd,et al.  Least squares quantization in PCM , 1982, IEEE Trans. Inf. Theory.

[4]  Richard S. Sutton,et al.  Neuronlike adaptive elements that can solve difficult learning control problems , 1983, IEEE Transactions on Systems, Man, and Cybernetics.

[5]  Peter J. Rousseeuw,et al.  Clustering by means of medoids , 1987 .

[6]  Laurence A. Wolsey,et al.  Integer and Combinatorial Optimization , 1988 .

[7]  Zbigniew Michalewicz,et al.  Genetic algorithms + data structures = evolution programs (2nd, extended ed.) , 1994 .

[8]  Andrew W. Moore,et al.  Memory-based Stochastic Optimization , 1995, NIPS.

[9]  Zbigniew Michalewicz,et al.  Genetic Algorithms + Data Structures = Evolution Programs , 1996, Springer Berlin Heidelberg.

[10]  Marco Locatelli,et al.  Bayesian Algorithms for One-Dimensional Global Optimization , 1997, J. Glob. Optim..

[11]  William J. Welch,et al.  Computer experiments and global optimization , 1997 .

[12]  Andrew W. Moore,et al.  A Nonparametric Approach to Noisy and Costly Optimization , 2000, ICML.

[13]  Donald R. Jones,et al.  A Taxonomy of Global Optimization Methods Based on Response Surfaces , 2001, J. Glob. Optim..

[14]  Victor P. Il'ev,et al.  An Approximation Guarantee of the Greedy Descent Algorithm for Minimizing a Supermodular Set Function , 2001, Discret. Appl. Math..

[15]  D. R. Bond,et al.  Electricity Production by Geobacter sulfurreducens Attached to Electrodes , 2003, Applied and Environmental Microbiology.

[16]  D. Park,et al.  Improved fuel cell and electrode designs for producing electricity from microbial degradation. , 2003, Biotechnology and bioengineering.

[17]  R. Battiti,et al.  A Memory-Based RASH Optimizer , 2006 .

[18]  Dale Schuurmans,et al.  Discriminative Batch Mode Active Learning , 2007, NIPS.

[19]  D. Lizotte Practical bayesian optimization , 2008 .

[20]  Xiaoli Z. Fern,et al.  Optimization of pH and nitrogen for enhanced hydrogen production by Synechocystis sp. PCC 6803 via statistical and machine learning methods , 2009, Biotechnology progress.

[21]  Carl E. Rasmussen,et al.  Gaussian processes for machine learning , 2005, Adaptive computation and machine learning.

[22]  Andrew M. Ross Computing Bounds on the Expected Maximum of Correlated Normal Variables , 2010 .

[23]  Nando de Freitas,et al.  A Tutorial on Bayesian Optimization of Expensive Cost Functions, with Application to Active User Modeling and Hierarchical Reinforcement Learning , 2010, ArXiv.