Scalable Bayesian optimization with high-dimensional outputs using randomized prior networks

Several fundamental problems in science and engineering consist of global optimization tasks involving unknown high-dimensional (black-box) functions that map a set of controllable variables to the outcomes of an expensive experiment. Bayesian Optimization (BO) techniques are known to be effective in tackling global optimization problems using a relatively small number objective function evaluations, but their performance suffers when dealing with high-dimensional outputs. To overcome the major challenge of dimensionality, here we propose a deep learning framework for BO and sequential decision making based on bootstrapped ensembles of neural architectures with randomized priors. Using appropriate architecture choices, we show that the proposed framework can approximate functional relationships between design variables and quantities of interest, even in cases where the latter take values in high-dimensional vector spaces or even infinite-dimensional function spaces. In the context of BO, we augmented the proposed probabilistic surrogates with re-parameterized Monte Carlo approximations of multiple-point (parallel) acquisition functions, as well as methodological extensions for accommodating black-box constraints and multi-fidelity information sources. We test the proposed framework against state-of-the-art methods for BO and demonstrate superior performance across several challenging tasks with high-dimensional outputs, including a constrained multi-fidelity optimization task involving shape optimization of rotor blades in turbo-machinery.

[1]  P. Perdikaris,et al.  Scalable Uncertainty Quantification for Deep Operator Networks using Randomized Priors , 2022, Computer Methods in Applied Mechanics and Engineering.

[2]  Wesley J. Maddox,et al.  Optimizing High-Dimensional Physics Simulations via Composite Bayesian Optimization , 2021, ArXiv.

[3]  Nikolaos Aletras,et al.  Active Learning by Acquiring Contrastive Examples , 2021, EMNLP.

[4]  Xuan Zeng,et al.  An Efficient Batch-Constrained Bayesian Optimization Approach for Analog Circuit Synthesis via Multiobjective Acquisition Ensemble , 2021, IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems.

[5]  Andrew Gordon Wilson,et al.  Bayesian Optimization with High-Dimensional Outputs , 2021, NeurIPS.

[6]  Robert W. Heath,et al.  Optimizing Coverage and Capacity in Cellular Networks using Machine Learning , 2020, ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[7]  Aditya Gopalan,et al.  No-regret Algorithms for Multi-task Bayesian Optimization , 2020, AISTATS.

[8]  P. Frazier,et al.  Bayesian Optimization of Risk Measures , 2020, NeurIPS.

[9]  Roman Garnett,et al.  Efficient Nonmyopic Bayesian Optimization via One-Shot Multi-Step Trees , 2020, NeurIPS.

[10]  Antoine Blanchard,et al.  Output-Weighted Importance Sampling for Bayesian Experimental Design and Uncertainty Quantification , 2020, ArXiv.

[11]  Jaime Fern'andez del R'io,et al.  Array programming with NumPy , 2020, Nature.

[12]  Alexander Ulanov,et al.  Interferobot: aligning an optical interferometer by a reinforcement learning agent , 2020, NeurIPS.

[13]  Maximilian Balandat,et al.  Differentiable Expected Hypervolume Improvement for Parallel Multi-Objective Bayesian Optimization , 2020, NeurIPS.

[14]  Antoine Blanchard,et al.  Informative Path Planning for Anomaly Detection in Environment Exploration and Monitoring , 2020, Ocean Engineering.

[15]  Richard E. Turner,et al.  Conservative Uncertainty Estimation By Fitting Prior Networks , 2020, ICLR.

[16]  Antoine Blanchard,et al.  Bayesian optimization with output-weighted optimal sampling , 2020, J. Comput. Phys..

[17]  David Zwicker,et al.  py-pde: A Python package for solving partial differential equations , 2020, J. Open Source Softw..

[18]  D. Bindel,et al.  Efficient Rollout Strategies for Bayesian Optimization , 2020, UAI.

[19]  Matthias Poloczek,et al.  Scalable Constrained Bayesian Optimization , 2020, AISTATS.

[20]  R. Calandra,et al.  Re-Examining Linear Embeddings for High-Dimensional Bayesian Optimization , 2020, NeurIPS.

[21]  Ryan-Rhys Griffiths,et al.  Constrained Bayesian optimization for automatic chemical design using variational autoencoders , 2019, Chemical science.

[22]  George Em Karniadakis,et al.  Learning nonlinear operators via DeepONet based on the universal approximation theorem of operators , 2019, Nature Machine Intelligence.

[23]  Matthias Poloczek,et al.  Scalable Global Optimization via Local Bayesian Optimization , 2019, NeurIPS.

[24]  Balaji Lakshminarayanan,et al.  Deep Ensembles: A Loss Landscape Perspective , 2019, ArXiv.

[25]  Roman Garnett,et al.  BINOCULARS for efficient, nonmyopic sequential experimental design , 2019, ICML.

[26]  Fan Zhang,et al.  Brief review of image denoising techniques , 2019, Vis. Comput. Ind. Biomed. Art.

[27]  Fan Zhang,et al.  Brief review of image denoising techniques , 2019, Visual Computing for Industry, Biomedicine, and Art.

[28]  Peter I. Frazier,et al.  Bayesian Optimization of Composite Functions , 2019, ICML.

[29]  Matthias Poloczek,et al.  A Framework for Bayesian Optimization in Embedded Subspaces , 2019, ICML.

[30]  Paris Perdikaris,et al.  Multi-fidelity classification using Gaussian processes: accelerating the prediction of large-scale computational models , 2019, Computer Methods in Applied Mechanics and Engineering.

[31]  Michael M. Joly,et al.  Machine Learning Enabled Adaptive Optimization of a Transonic Compressor Rotor With Precompression , 2019, Journal of Turbomachinery.

[32]  Anders Kirk Uhrenholt,et al.  Efficient Bayesian Optimization for Target Vector Estimation , 2019, AISTATS.

[33]  K. S. Sesh Kumar,et al.  High-Dimensional Bayesian Optimization with Manifold Gaussian Processes , 2019, ArXiv.

[34]  Victor M. Becerra,et al.  Optimal control , 2008, Scholarpedia.

[35]  Dirk Husmeier,et al.  On a New Improvement-Based Acquisition Function for Bayesian Optimization , 2018, ArXiv.

[36]  Alex Sherstinsky,et al.  Fundamentals of Recurrent Neural Network (RNN) and Long Short-Term Memory (LSTM) Network , 2018, Physica D: Nonlinear Phenomena.

[37]  R. Tütüncü,et al.  Optimization Methods in Finance , 2018 .

[38]  Neil D. Lawrence,et al.  Structured Variationally Auto-encoded Optimization , 2018, ICML.

[39]  Albin Cassirer,et al.  Randomized Prior Functions for Deep Reinforcement Learning , 2018, NeurIPS.

[40]  Frank Hutter,et al.  Maximizing acquisition functions for Bayesian optimization , 2018, NeurIPS.

[41]  Andrew Gordon Wilson,et al.  Averaging Weights Leads to Wider Optima and Better Generalization , 2018, UAI.

[42]  Matthias Poloczek,et al.  Advances in Bayesian Optimization with Applications in Aerospace Engineering , 2018 .

[43]  Benjamin Van Roy,et al.  A Tutorial on Thompson Sampling , 2017, Found. Trends Mach. Learn..

[44]  Roman Garnett,et al.  Discovering and Exploiting Additive Structure for Bayesian Optimization , 2017, AISTATS.

[45]  Matthias Poloczek,et al.  Bayesian Optimization with Gradients , 2017, NIPS.

[46]  Zi Wang,et al.  Batched High-dimensional Bayesian Optimization via Structural Kernel Learning , 2017, ICML.

[47]  Charles Blundell,et al.  Simple and Scalable Predictive Uncertainty Estimation using Deep Ensembles , 2016, NIPS.

[48]  Peter I. Frazier,et al.  The Parallel Knowledge Gradient Method for Batch Bayesian Optimization , 2016, NIPS.

[49]  Anthony N. Pettitt,et al.  A Review of Modern Computational Algorithms for Bayesian Optimal Design , 2016 .

[50]  Matthias Poloczek,et al.  Multi-Information Source Optimization , 2016, NIPS.

[51]  Keiron O'Shea,et al.  An Introduction to Convolutional Neural Networks , 2015, ArXiv.

[52]  Kirthevasan Kandasamy,et al.  High Dimensional Bayesian Optimisation and Bandits via Additive Models , 2015, ICML.

[53]  Jasper Snoek,et al.  Multi-Task Bayesian Optimization , 2013, NIPS.

[54]  Nicolas Vayatis,et al.  Parallel Gaussian Process Optimization with Upper Confidence Bound and Pure Exploration , 2013, ECML/PKDD.

[55]  David Ginsbourger,et al.  Fast Computation of the Multi-Points Expected Improvement with Applications in Batch Selection , 2013, LION.

[56]  Andreas Krause,et al.  Parallelizing Exploration-Exploitation Tradeoffs with Gaussian Process Bandit Optimization , 2012, ICML.

[57]  Andreas Krause,et al.  Contextual Gaussian Process Bandit Optimization , 2011, NIPS.

[58]  Carl E. Rasmussen,et al.  Additive Gaussian Processes , 2011, NIPS.

[59]  Philipp Hennig,et al.  Entropy Search for Information-Efficient Global Optimization , 2011, J. Mach. Learn. Res..

[60]  Michael T. M. Emmerich,et al.  Hypervolume-based expected improvement: Monotonicity properties and exact computation , 2011, 2011 IEEE Congress of Evolutionary Computation (CEC).

[61]  Alan Fern,et al.  Batch Bayesian Optimization via Simulation Matching , 2010, NIPS.

[62]  R. Haftka,et al.  Surrogate-based Optimization with Parallel Simulations using the Probability of Improvement , 2010 .

[63]  Yoshua Bengio,et al.  Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.

[64]  Andreas Krause,et al.  Information-Theoretic Regret Bounds for Gaussian Process Optimization in the Bandit Setting , 2009, IEEE Transactions on Information Theory.

[65]  Fei-Fei Li,et al.  ImageNet: A large-scale hierarchical image database , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.

[66]  Stefan M. Wild,et al.  Bayesian Calibration and Uncertainty Analysis for Computationally Expensive Models Using Optimization and Radial Basis Function Approximation , 2008 .

[67]  Warren B. Powell,et al.  A Knowledge-Gradient Policy for Sequential Information Collection , 2008, SIAM J. Control. Optim..

[68]  John D. Hunter,et al.  Matplotlib: A 2D Graphics Environment , 2007, Computing in Science & Engineering.

[69]  Michael T. M. Emmerich,et al.  Single- and multiobjective evolutionary optimization assisted by Gaussian random field metamodels , 2006, IEEE Transactions on Evolutionary Computation.

[70]  Joshua D. Knowles,et al.  ParEGO: a hybrid algorithm with on-line landscape approximation for expensive multiobjective optimization problems , 2006, IEEE Transactions on Evolutionary Computation.

[71]  David E. Goldberg,et al.  Multi-objective bayesian optimization algorithm , 2002 .

[72]  L. Breiman Random Forests , 2001, Encyclopedia of Machine Learning and Data Mining.

[73]  Charles Audet,et al.  A surrogate-model-based method for constrained optimization , 2000 .

[74]  S. Hochreiter,et al.  Long Short-Term Memory , 1997, Neural Computation.

[75]  J. Ross Quinlan,et al.  Induction of Decision Trees , 1986, Machine Learning.

[76]  R. M. Hicks,et al.  Wing Design by Numerical Optimization , 1977 .

[77]  Harold J. Kushner,et al.  A New Method of Locating the Maximum Point of an Arbitrary Multipeak Curve in the Presence of Noise , 1964 .

[78]  Seyed Mohammad Asghari,et al.  Evaluating Predictive Distributions: Does Bayesian Deep Learning Work? , 2021, ArXiv.

[79]  B. Feng,et al.  Thinking Inside the Box: A Tutorial on Grey-Box Bayesian Optimization , 2021, 2021 Winter Simulation Conference (WSC).

[80]  Daniel R. Jiang,et al.  BoTorch: A Framework for Efficient Monte-Carlo Bayesian Optimization , 2020, NeurIPS.

[81]  Hongzi Mao,et al.  High-Dimensional Contextual Policy Search with Unknown Context Rewards using Bayesian Optimization , 2020, NeurIPS.

[82]  Jian Wu,et al.  Practical Two-Step Lookahead Bayesian Optimization , 2019, NeurIPS.

[83]  Kirthevasan Kandasamy,et al.  Offline Contextual Bayesian Optimization , 2019, NeurIPS.

[84]  Andreas Krause,et al.  Efficient High Dimensional Bayesian Optimization with Additivity and Quadrature Fourier Features , 2018, NeurIPS.

[85]  Nando de Freitas,et al.  Taking the Human Out of the Loop: A Review of Bayesian Optimization , 2016, Proceedings of the IEEE.

[86]  D. Ginsbourger,et al.  Kriging is well-suited to parallelize optimization , 2010 .

[87]  Andy J. Keane,et al.  Recent advances in surrogate-based optimization , 2009 .

[88]  Raphael T. Haftka,et al.  Surrogate-based Analysis and Optimization , 2005 .

[89]  Raino A. E. Mäkinen,et al.  Introduction to shape optimization - theory, approximation, and computation , 2003, Advances in design and control.