Honest exploration of intractable probability distributions via Markov chain Monte Carlo

被引:164
|
作者
Jones, GL
Hobert, JP
机构
[1] Univ Minnesota, Sch Stat, Minneapolis, MN 55455 USA
[2] Univ Florida, Dept Stat, Gainesville, FL 32611 USA
关键词
central limit theorem; convergence rate; coupling inequality; drift condition; general state space; geometric ergodicity; Gibbs sampler; hierarchical random effects model; Metropolis algorithm; minorization condition; regeneration; splitting; uniform ergodicity;
D O I
10.1214/ss/1015346317
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
Two important questions that must be answered whenever a Markov chain Monte Carlo (MCMC) algorithm is used are (Q1) What is an appropriate burn-in? and (Q2) How long should the sampling continue after burn-in? Developing rigorous answers to these questions presently requires a detailed study of the convergence properties of the underlying Markov chain. Consequently, in most practical applications of MCMC, exact answers to (Q1) and (Q2) are not sought. The goal of this paper is to demystify the analysis that leads to honest answers to (Q1) and (Q2). The authors hope that this article will serve as a bridge between those developing Markov chain theory and practitioners using MCMC to solve practical problems. The ability to address (Q1) and (Q2) formally comes from establishing a drift condition and an associated minorization condition, which together imply that the underlying Markov chain is geometrically ergodic. In this article, we explain exactly what drift and minorization are as well as how and why these conditions can be used to form rigorous answers to (Q1) and (Q2). The basic ideas are as follows. The results of Rosenthal (1995) and Roberts and Tweedie (1999) allow one to use drift and minorization conditions to construct a formula giving an analytic upper bound on the distance to stationarity, A rigorous answer to (Q1) can be calculated using this formula. The desired characteristics of the target distribution are typically estimated using ergodic averages. Geometric ergodicity of the underlying Markov chain implies that there are central limit theorems available for ergodic averages (Chan and Geyer 1994). The regenerative simulation technique (Mykland, Tierney and Yu, 1995; Robert, 1995) can be used to get a consistent estimate of the variance of the asymptotic normal distribution. Hence, an asymptotic standard error can be calculated, which provides an answer to (Q2) in the sense that an appropriate time to stop sampling can be determined. The methods are illustrated using a Gibbs sampler for a Bayesian version of the one-way random effects model and a data set concerning styrene exposure.
引用
收藏
页码:312 / 334
页数:23
相关论文
共 50 条
  • [1] Unbiased Markov chain Monte Carlo for intractable target distributions
    Middleton, Lawrence
    Deligiannidis, George
    Doucet, Arnaud
    Jacob, Pierre E.
    ELECTRONIC JOURNAL OF STATISTICS, 2020, 14 (02): : 2842 - 2891
  • [2] An efficient Markov chain Monte Carlo method for distributions with intractable normalising constants
    Moller, J.
    Pettitt, A. N.
    Reeves, R.
    Berthelsen, K. K.
    BIOMETRIKA, 2006, 93 (02) : 451 - 458
  • [3] Estimation via Markov chain Monte Carlo
    Spall, JC
    IEEE CONTROL SYSTEMS MAGAZINE, 2003, 23 (02): : 34 - 45
  • [4] Estimation via Markov chain Monte Carlo
    Spall, JC
    PROCEEDINGS OF THE 2002 AMERICAN CONTROL CONFERENCE, VOLS 1-6, 2002, 1-6 : 2559 - 2564
  • [5] Efficient Bernoulli factory Markov chain Monte Carlo for intractable posteriors
    Vats, D.
    Goncalves, F. B.
    Latuszynski, K.
    Roberts, G. O.
    BIOMETRIKA, 2022, 109 (02) : 369 - 385
  • [6] Markov Chain Monte Carlo With Mixtures of Mutually Singular Distributions
    Gottardo, Raphael
    Raftery, Adrian E.
    JOURNAL OF COMPUTATIONAL AND GRAPHICAL STATISTICS, 2008, 17 (04) : 949 - 975
  • [7] Bayesian Computation Via Markov Chain Monte Carlo
    Craiu, Radu V.
    Rosenthal, Jeffrey S.
    ANNUAL REVIEW OF STATISTICS AND ITS APPLICATION, VOL 1, 2014, 1 : 179 - 201
  • [8] Sampling from complicated and unknown distributions Monte Carlo and Markov Chain Monte Carlo methods for redistricting
    Cho, Wendy K. Tam
    Liu, Yan Y.
    PHYSICA A-STATISTICAL MECHANICS AND ITS APPLICATIONS, 2018, 506 : 170 - 178
  • [9] Markov Chain Monte Carlo
    Henry, Ronnie
    EMERGING INFECTIOUS DISEASES, 2019, 25 (12) : 2298 - 2298
  • [10] Parallel Markov Chain Monte Carlo via Spectral Clustering
    Basse, Guillaume
    Pillai, Natesh
    Smith, Aaron
    ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 51, 2016, 51 : 1318 - 1327