Efficient nonmyopic batch active search

Active search is a learning paradigm for actively identifying as many members of a given class as possible. A critical target scenario is high-throughput screening for scientific discovery, such as drug or materials discovery. In these settings, specialized instruments can often evaluate \emph{multiple} points simultaneously; however, all existing work on active search focuses on sequential acquisition. We bridge this gap, addressing batch active search from both the theoretical and practical perspective. We first derive the Bayesian optimal policy for this problem, then prove a lower bound on the performance gap between sequential and batch optimal policies: the ``cost of parallelization.'' We also propose novel, efficient batch policies inspired by state-of-the-art sequential policies, and develop an aggressive pruning technique that can dramatically speed up computation. We conduct thorough experiments on data from three application domains: a citation network, material science, and drug discovery, testing all proposed policies (14 total) with a wide range of batch sizes. Our results demonstrate that the empirical performance gap matches our theoretical bound, that nonmyopic policies usually significantly outperform myopic alternatives, and that diversity is an important consideration for batch policy design.

[1]  Ross D. King,et al.  Cheaper faster drug development validated by the repositioning of drugs against neglected tropical diseases , 2015, Journal of The Royal Society Interface.

[2]  Harold J. Kushner,et al.  A New Method of Locating the Maximum Point of an Arbitrary Multipeak Curve in the Presence of Noise , 1964 .

[3]  Peter I. Frazier,et al.  The Parallel Knowledge Gradient Method for Batch Bayesian Optimization , 2016, NIPS.

[4]  Rong Jin,et al.  Batch mode active learning and its application to medical image classification , 2006, ICML.

[5]  Gunnar Rätsch,et al.  Active Learning with Support Vector Machines in the Drug Discovery Process , 2003, J. Chem. Inf. Comput. Sci..

[6]  John J. Irwin,et al.  ZINC 15 – Ligand Discovery for Everyone , 2015, J. Chem. Inf. Model..

[7]  Andreas Krause,et al.  Discovering Valuable items from Massive Data , 2015, KDD.

[8]  Robert D. Nowak,et al.  Top Arm Identification in Multi-Armed Bandits with Batch Arm Pulls , 2016, AISTATS.

[9]  Dale Schuurmans,et al.  Discriminative Batch Mode Active Learning , 2007, NIPS.

[10]  Andreas Krause,et al.  Near-optimal Batch Mode Active Learning and Adaptive Submodular Optimization , 2013, ICML.

[11]  Roman Garnett,et al.  Efficient Nonmyopic Active Search , 2017, ICML.

[12]  Roman Garnett,et al.  Active Search for Computer‐aided Drug Design , 2018, Molecular informatics.

[13]  Alan Fern,et al.  Batch Bayesian Optimization via Simulation Matching , 2010, NIPS.

[14]  Pushmeet Kohli,et al.  Batched Gaussian Process Bandit Optimization via Determinantal Point Processes , 2016, NIPS.

[15]  Roman Garnett,et al.  Introducing the ‘active search’ method for iterative virtual screening , 2015, Journal of Computer-Aided Molecular Design.

[16]  Jialei Wang,et al.  Bayesian Optimization with Parallel Function Evaluations and Multiple Information Sources: Methodology with Applications in Biochemistry, Aerospace Engineering, and Machine Learning , 2017 .

[17]  D. Ginsbourger,et al.  A Multi-points Criterion for Deterministic Parallel Global Optimization based on Gaussian Processes , 2008 .

[18]  Sethuraman Panchanathan,et al.  Adaptive Batch Mode Active Learning , 2015, IEEE Transactions on Neural Networks and Learning Systems.

[19]  Jeff G. Schneider,et al.  Active Search and Bandits on Graphs using Sigma-Optimality , 2015, UAI.

[20]  J. Schneider,et al.  Bayesian Optimal Active Search on Graphs , 2011 .

[21]  M. L. Fisher,et al.  An analysis of approximations for maximizing submodular set functions—I , 1978, Math. Program..

[22]  Barnabás Póczos,et al.  Active learning and search on low-rank matrices , 2013, KDD.

[23]  Amin Saberi,et al.  Stochastic Submodular Maximization , 2008, WINE.

[24]  Andreas Krause,et al.  Information-Theoretic Regret Bounds for Gaussian Process Optimization in the Bandit Setting , 2009, IEEE Transactions on Information Theory.

[25]  Roman Garnett,et al.  Bayesian Optimal Active Search and Surveying , 2012, ICML.

[26]  Neil D. Lawrence,et al.  Batch Bayesian Optimization via Local Penalization , 2015, AISTATS.

[27]  Roman Garnett,et al.  Active search on graphs , 2013, KDD.

[28]  Gunnar Rätsch,et al.  Active Learning in the Drug Discovery Process , 2001, NIPS.

[29]  Zoubin Ghahramani,et al.  Parallel Predictive Entropy Search for Batch Global Optimization of Expensive Objective Functions , 2015, NIPS.

[30]  D. Ginsbourger,et al.  Kriging is well-suited to parallelize optimization , 2010 .

[31]  Nan Ye,et al.  Active Learning for Probabilistic Hypotheses Using the Maximum Gibbs Error Criterion , 2013, NIPS.

[32]  Alán Aspuru-Guzik,et al.  Parallel and Distributed Thompson Sampling for Large-scale Accelerated Exploration of Chemical Space , 2017, ICML.

[33]  Roman Garnett,et al.  Active Search in Intensionally Specified Structured Spaces , 2017, AAAI.

[34]  Andreas Krause,et al.  Parallelizing Exploration-Exploitation Tradeoffs with Gaussian Process Bandit Optimization , 2012, ICML.

[35]  Roman Garnett,et al.  Active Pointillistic Pattern Search , 2015, AISTATS.