A fast asynchronous Markov chain Monte Carlo sampler for sparse Bayesian inference

被引:0
作者
Atchade, Yves [1 ,2 ]
Wang, Liwei [1 ]
机构
[1] Boston Univ, Dept Math & Stat, Boston, MA USA
[2] Boston Univ, Dept Math & Stat, 665 Commonwealth Ave, Boston, MA 02215 USA
关键词
asynchronous MCMC sampling; Bayesian deep learning; MCMC mixing; sparse Bayesian inference; 62Jxx; SELECTION; REGRESSION; HORSESHOE;
D O I
10.1093/jrsssb/qkad078
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
We propose a very fast approximate Markov chain Monte Carlo sampling framework that is applicable to a large class of sparse Bayesian inference problems. The computational cost per iteration in several regression models is of order O(n(s+J)), where n is the sample size, s is the underlying sparsity of the model, and J is the size of a randomly selected subset of regressors. This cost can be further reduced by data sub-sampling when stochastic gradient Langevin dynamics are employed. The algorithm is an extension of the asynchronous Gibbs sampler of Johnson et al. [(2013). Analyzing Hogwild parallel Gaussian Gibbs sampling. In Proceedings of the 26th International Conference on Neural Information Processing Systems (NIPS'13) (Vol. 2, pp. 2715-2723)], but can be viewed from a statistical perspective as a form of Bayesian iterated sure independent screening [Fan, J., Samworth, R., & Wu, Y. (2009). Ultrahigh dimensional feature selection: Beyond the linear model. Journal of Machine Learning Research, 10, 2013-2038]. We show that in high-dimensional linear regression problems, the Markov chain generated by the proposed algorithm admits an invariant distribution that recovers correctly the main signal with high probability under some statistical assumptions. Furthermore, we show that its mixing time is at most linear in the number of regressors. We illustrate the algorithm with several models.
引用
收藏
页码:1492 / 1516
页数:25
相关论文
共 55 条
[1]  
[Anonymous], 2013, ADV NEURAL INFORM PR
[2]  
Atchade Y., 2019, ARXIV
[3]  
Atchade Y., 2011, Bayesian Time Series Models, P32
[4]   ON THE CONTRACTION PROPERTIES OF SOME HIGH-DIMENSIONAL QUASI-POSTERIOR DISTRIBUTIONS [J].
Atchade, Yves A. .
ANNALS OF STATISTICS, 2017, 45 (05) :2248-2273
[5]   Approximate Spectral Gaps for Markov Chain Mixing Times in High Dimensions [J].
Atchade, Yves F. .
SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 2021, 3 (03) :854-872
[6]   Horseshoe Regularisation for Machine Learning in Complex and Deep Models [J].
Bhadra, Anindya ;
Datta, Jyotishka ;
Li, Yunfan ;
Polson, Nicholas G. .
INTERNATIONAL STATISTICAL REVIEW, 2020, 88 (02) :302-320
[7]   Fast sampling with Gaussian scale mixture priors in high-dimensional regression [J].
Bhattacharya, Anirban ;
Chakraborty, Antik ;
Mallick, Bani K. .
BIOMETRIKA, 2016, 103 (04) :985-991
[8]  
Biswas N., 2021, ARXIV
[9]  
Biswas N., 2019, ARXIV
[10]  
Blei D.M., 2016, arXiv