TI-Stan: Model Comparison Using Thermodynamic Integration and HMC

We present a novel implementation of the adaptively annealed thermodynamic integration technique using Hamiltonian Monte Carlo (HMC). Thermodynamic integration with importance sampling and adaptive annealing is an especially useful method for estimating model evidence for problems that use physics-based mathematical models. Because it is based on importance sampling, this method requires an efficient way to refresh the ensemble of samples. Existing successful implementations use binary slice sampling on the Hilbert curve to accomplish this task. This implementation works well if the model has few parameters or if it can be broken into separate parts with identical parameter priors that can be refreshed separately. However, for models that are not separable and have many parameters, a different method for refreshing the samples is needed. HMC, in the form of the MC-Stan package, is effective for jointly refreshing the ensemble under a high-dimensional model. MC-Stan uses automatic differentiation to compute the gradients of the likelihood that HMC requires in about the same amount of time as it computes the likelihood function itself, easing the programming burden compared to implementations of HMC that require explicitly specified gradient functions. We present a description of the overall TI-Stan procedure and results for representative example problems.

[1]  Jiqiang Guo,et al.  Stan: A Probabilistic Programming Language. , 2017, Journal of statistical software.

[2]  J. Skilling Nested sampling for general Bayesian computation , 2006 .

[3]  Radford M. Neal Slice Sampling , 2003, The Annals of Statistics.

[4]  Robert Wesley Henderson,et al.  Design and Analysis of Efficient Parallel Bayesian Model Comparison Algorithms , 2019 .

[5]  G. L. Bretthorst Nonuniform sampling: Bandwidth and aliasing , 2001 .

[6]  Radford M. Neal MCMC Using Hamiltonian Dynamics , 2011, 1206.1901.

[7]  Djc MacKay,et al.  Slice sampling - Discussion , 2003 .

[8]  Xiao-Li Meng,et al.  Simulating Normalizing Constants: From Importance Sampling to Bridge Sampling to Path Sampling , 1998 .

[9]  Wolfgang von der Linden,et al.  Bayesian Probability Theory: Applications in the Physical Sciences , 2014 .

[10]  J. Skilling Galilean and Hamiltonian Monte Carlo , 2019 .

[11]  Mark Girolami,et al.  The Controlled Thermodynamic Integral for Bayesian Model Evidence Evaluation , 2016 .

[12]  Andrew Gelman,et al.  The No-U-turn sampler: adaptively setting path lengths in Hamiltonian Monte Carlo , 2011, J. Mach. Learn. Res..

[13]  J. Kirkwood Statistical Mechanics of Fluid Mixtures , 1935 .

[14]  Using the Z-Order Curve for Bayesian Model Comparison , 2017 .

[15]  Rong Chen,et al.  A Theoretical Framework for Sequential Importance Sampling with Resampling , 2001, Sequential Monte Carlo Methods in Practice.

[16]  A. Lasenby,et al.  polychord: next-generation nested sampling , 2015, 1506.00171.

[17]  P. Goggans,et al.  TI-Stan: Adaptively Annealed Thermodynamic Integration with HMC † , 2019 .

[18]  Marvin H. J. Guber Bayesian Spectrum Analysis and Parameter Estimation , 1988 .

[19]  James M. Kang,et al.  Space-Filling Curves , 2017, Encyclopedia of GIS.

[20]  S. Walker Invited comment on the paper "Slice Sampling" by Radford Neal , 2003 .

[21]  John Skilling Programming the Hilbert curve , 2004 .

[22]  P. Goggans,et al.  Using Thermodynamic Integration to Calculate the Posterior Probability in Bayesian Model Selection Problems , 2004 .

[23]  Dominicus Kester,et al.  BAYESIAN INFERENCE AND MAXIMUM ENTROPY METHODS IN SCIENCE AND ENGINEERING , 2010 .

[24]  F. Feroz,et al.  MultiNest: an efficient and robust Bayesian inference tool for cosmology and particle physics , 2008, 0809.3437.