Confidence Sets and Hypothesis Testing in a Likelihood-Free Inference Setting

Parameter estimation, statistical tests and confidence sets are the cornerstones of classical statistics that allow scientists to make inferences about the underlying process that generated the observed data. A key question is whether one can still construct hypothesis tests and confidence sets with proper coverage and high power in a so-called likelihood-free inference (LFI) setting; that is, a setting where the likelihood is not explicitly known but one can forward-simulate observable data according to a stochastic model. In this paper, we present $\texttt{ACORE}$ (Approximate Computation via Odds Ratio Estimation), a frequentist approach to LFI that first formulates the classical likelihood ratio test (LRT) as a parametrized classification problem, and then uses the equivalence of tests and confidence sets to build confidence regions for parameters of interest. We also present a goodness-of-fit procedure for checking whether the constructed tests and confidence regions are valid. $\texttt{ACORE}$ is based on the key observation that the LRT statistic, the rejection probability of the test, and the coverage of the confidence set are conditional distribution functions which often vary smoothly as a function of the parameters of interest. Hence, instead of relying solely on samples simulated at fixed parameter settings (as is the convention in standard Monte Carlo solutions), one can leverage machine learning tools and data simulated in the neighborhood of a parameter to improve estimates of quantities of interest. We demonstrate the efficacy of $\texttt{ACORE}$ with both theoretical and empirical results. Our implementation is available on Github.

[1]  Shigeo Abe DrEng Pattern Classification , 2001, Springer London.

[2]  Tom Charnock,et al.  Fast likelihood-free cosmology with neural density estimators and active learning , 2019, Monthly Notices of the Royal Astronomical Society.

[3]  Jakob H. Macke,et al.  Likelihood-free inference with emulator networks , 2018, AABI.

[4]  Iain Murray,et al.  Sequential Neural Likelihood: Fast Likelihood-free Inference with Autoregressive Flows , 2018, AISTATS.

[5]  R. Cousins,et al.  Lectures on Statistics in Theory: Prelude to Statistics in Practice. , 2018, 1807.05996.

[6]  D. Whiteson,et al.  Deep Learning and Its Application to LHC Physics , 2018, Annual Review of Nuclear and Particle Science.

[7]  R. Cousins,et al.  A Unified Approach to the Classical Statistical Analysis of Small Signals , 1997, physics/9711021.

[8]  David J. C. MacKay,et al.  Information Theory, Inference, and Learning Algorithms , 2004, IEEE Transactions on Information Theory.

[9]  Gerda Claeskens,et al.  Bootstrap confidence bands for regression curves and their derivatives , 2003 .

[10]  Rafael Izbicki,et al.  Validation of Approximate Likelihood and Emulator Models for Computationally Intensive Simulations , 2019, AISTATS.

[11]  Christopher C. Drovandi,et al.  Variational Bayes with synthetic likelihood , 2016, Statistics and Computing.

[12]  Gerda Claeskens,et al.  Simultaneous Confidence Bands for Penalized Spline Estimators , 2009 .

[13]  P. Diggle,et al.  Monte Carlo Methods of Inference for Implicit Statistical Models , 1984 .

[14]  Shakir Mohamed,et al.  Learning in Implicit Generative Models , 2016, ArXiv.

[15]  Xin Fu,et al.  Confidence bands in nonparametric regression , 2009 .

[16]  John A. Nelder,et al.  A Simplex Method for Function Minimization , 1965, Comput. J..

[17]  David S. Greenberg,et al.  Automatic Posterior Transformation for Likelihood-Free Inference , 2019, ICML.

[18]  Donald B. Rubin,et al.  Validation of Software for Bayesian Models Using Posterior Quantiles , 2006 .

[19]  Michael U. Gutmann,et al.  Adaptive Gaussian Copula ABC , 2019, AISTATS.

[20]  Natalia Gimelshein,et al.  PyTorch: An Imperative Style, High-Performance Deep Learning Library , 2019, NeurIPS.

[21]  Gaël Varoquaux,et al.  Scikit-learn: Machine Learning in Python , 2011, J. Mach. Learn. Res..

[22]  Min‐ge Xie,et al.  Approximate confidence distribution computing: An effective likelihood-free method with statistical guarantees , 2017, 1705.10347.

[23]  Nicolai Meinshausen,et al.  Quantile Regression Forests , 2006, J. Mach. Learn. Res..

[24]  William F. Podlaski,et al.  Training deep neural density estimators to identify mechanistic models of neural dynamics , 2020, eLife.

[25]  Ann B. Lee,et al.  ABC–CDE: Toward Approximate Bayesian Computation With Complex High-Dimensional Data and Limited Simulations , 2018, Journal of Computational and Graphical Statistics.

[26]  Aki Vehtari,et al.  Validating Bayesian Inference Algorithms with Simulation-Based Calibration , 2018, 1804.06788.

[27]  David T. Frazier,et al.  Bayesian Synthetic Likelihood , 2017, 2305.05120.

[28]  Jean-Michel Marin,et al.  Approximate Bayesian computational methods , 2011, Statistics and Computing.

[29]  Philip B. Stark,et al.  Constructing Confidence Regions of Optimal Expected Size , 2009 .

[30]  Yanan Fan,et al.  Handbook of Approximate Bayesian Computation , 2018 .

[31]  .. W. V. Der,et al.  On Profile Likelihood , 2000 .

[32]  Kyle Cranmer,et al.  Practical Statistics for the LHC , 2014, 1503.07622.

[33]  K. Cranmer,et al.  MadMiner: Machine Learning-Based Inference for Particle Physics , 2019, Computing and Software for Big Science.

[34]  Gilles Louppe,et al.  Likelihood-free MCMC with Amortized Approximate Likelihood Ratios , 2019 .

[35]  Gilles Louppe,et al.  The frontier of simulation-based inference , 2020, Proceedings of the National Academy of Sciences.

[36]  D. Balding,et al.  Approximate Bayesian computation in population genetics. , 2002, Genetics.

[37]  Jasper Snoek,et al.  Practical Bayesian Optimization of Machine Learning Algorithms , 2012, NIPS.

[38]  The LSST Dark Energy Science Collaboration,et al.  Evaluation of probabilistic photometric redshift estimation approaches for LSST , 2020 .

[39]  H. Katzgraber Introduction to Monte Carlo Methods , 2009, 0905.1629.

[40]  Pierre Baldi,et al.  Parameterized neural networks for high-energy physics , 2016, The European Physical Journal C.

[41]  K. Perez Nuclear Instruments and Methods in Physics Research Section A: Accelerators, Spectrometers, Detectors and Associated Equipment , 2014 .

[42]  Alexander Vandenberg-Rodes,et al.  Modeling Smooth Backgrounds and Generic Localized Signals with Gaussian Processes , 2017, 1709.05681.

[43]  Wolfgang A. Rolke,et al.  Limits and confidence intervals in the presence of nuisance parameters , 2004, physics/0403059.

[44]  Gilles Louppe,et al.  Mining gold from implicit models to improve likelihood-free inference , 2018, Proceedings of the National Academy of Sciences.

[45]  Yuhong Yang,et al.  Information Theory, Inference, and Learning Algorithms , 2005 .

[46]  Roger Barlow,et al.  Fitting using finite Monte Carlo samples , 1993 .

[47]  Gilles Louppe,et al.  Approximating Likelihood Ratios with Calibrated Discriminative Classifiers , 2015, 1506.02169.

[48]  Jayanta K. Ghosh,et al.  On the Relation Among Shortest Confidence Intervals of Different Types , 1961 .

[49]  J. T. Childers,et al.  Observation of a new particle in the search for the Standard Model Higgs boson with the ATLAS detector at the LHC , 2012 .

[50]  Rafael Izbicki,et al.  High-Dimensional Density Ratio Estimation with Extensions to Approximate Likelihood Computation , 2014, AISTATS.

[51]  Simon J. Lilly,et al.  Photo‐z performance for precision cosmology , 2009, 0910.5735.

[52]  Jukka Corander,et al.  Likelihood-Free Inference by Ratio Estimation , 2016, Bayesian Analysis.

[53]  J. Pratt Length of Confidence Intervals , 1961 .

[54]  J. Neyman Outline of a Theory of Statistical Estimation Based on the Classical Theory of Probability , 1937 .

[55]  Florent Leclercq,et al.  Bayesian optimization for likelihood-free cosmological inference , 2018, Physical Review D.

[56]  Jean-Michel Marin,et al.  ABC random forests for Bayesian parameter inference , 2019, Bioinform..

[57]  T. Lai,et al.  Hybrid resampling methods for confidence intervals , 1998 .

[58]  David S. Greenberg,et al.  Training deep neural density estimators to identify mechanistic models of neural dynamics , 2019, bioRxiv.

[59]  Michael Woodroofe,et al.  ON THE UNIFIED METHOD WITH NUISANCE PARAMETERS , 2009 .

[60]  Ritabrata Dutta,et al.  Likelihood-free inference via classification , 2014, Stat. Comput..

[61]  Michael U. Gutmann,et al.  Dynamic Likelihood-free Inference via Ratio Estimation (DIRE) , 2018, ArXiv.