Adaptive Markov Chain Monte Carlo through Regeneration

Abstract Markov chain Monte Carlo (MCMC) is used for evaluating expectations of functions of interest under a target distribution π. This is done by calculating averages over the sample path of a Markov chain having π as its stationary distribution. For computational efficiency, the Markov chain should be rapidly mixing. This sometimes can be achieved only by careful design of the transition kernel of the chain, on the basis of a detailed preliminary exploratory analysis of π. An alternative approach might be to allow the transition kernel to adapt whenever new features of π are encountered during the MCMC run. However, if such adaptation occurs infinitely often, then the stationary distribution of the chain may be disturbed. We describe a framework, based on the concept of Markov chain regeneration, which allows adaptation to occur infinitely often but does not disturb the stationary distribution of the chain or the consistency of sample path averages.

[1]  N. Metropolis,et al.  Equation of State Calculations by Fast Computing Machines , 1953, Resonance.

[2]  W. K. Hastings,et al.  Monte Carlo Sampling Methods Using Markov Chains and Their Applications , 1970 .

[3]  Donald Geman,et al.  Stochastic Relaxation, Gibbs Distributions, and the Bayesian Restoration of Images , 1984, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[4]  E. Nummelin General irreducible Markov chains and non-negative operators: List of symbols and notation , 1984 .

[5]  E. Nummelin General irreducible Markov chains and non-negative operators: Embedded renewal processes , 1984 .

[6]  James R Murphy,et al.  Nonlinear Regression Modelling , 1985 .

[7]  Douglas M. Bates,et al.  Nonlinear Regression Analysis and Its Applications , 1988 .

[8]  Adrian F. M. Smith,et al.  Sampling-Based Approaches to Calculating Marginal Densities , 1990 .

[9]  R. Durrett Probability: Theory and Examples , 1993 .

[10]  G. Milliken Nonlinear Regression Analysis and Its Applications , 1990 .

[11]  Bradley P. Carlin,et al.  An iterative Monte Carlo method for nonconjugate Bayesian analysis , 1991 .

[12]  M. Newton Approximate Bayesian-inference With the Weighted Likelihood Bootstrap , 1994 .

[13]  Walter R. Gilks,et al.  Adaptive Direction Sampling , 1994 .

[14]  L. Tierney Markov Chains for Exploring Posterior Distributions , 1994 .

[15]  A. Gelfand,et al.  On Markov Chain Monte Carlo Acceleration , 1994 .

[16]  Kerrie Mengersen,et al.  [Bayesian Computation and Stochastic Systems]: Rejoinder , 1995 .

[17]  A. Gelfand,et al.  Efficient parametrisations for normal linear mixed models , 1995 .

[18]  J. Besag,et al.  Bayesian Computation and Stochastic Systems , 1995 .

[19]  Bin Yu,et al.  Regeneration in Markov chain samplers , 1995 .

[20]  Sylvia Richardson,et al.  Markov chain concepts related to sampling algorithms , 1995 .

[21]  C. Robert Convergence Control Methods for Markov Chain Monte Carlo Algorithms , 1995 .

[22]  Walter R. Gilks,et al.  Strategies for improving MCMC , 1995 .

[23]  R. Tweedie,et al.  Geometric convergence and central limit theorems for multidimensional Hastings and Metropolis algorithms , 1996 .

[24]  Sylvia Richardson,et al.  Markov Chain Monte Carlo in Practice , 1997 .

[25]  Mary Kathryn Cowles,et al.  A simulation approach to convergence rates for Markov chain Monte Carlo algorithms , 1998, Stat. Comput..