Proximal Markov chain Monte Carlo algorithms

被引:0
作者
Marcelo Pereyra
机构
[1] University of Bristol,Department of Mathematics
来源
Statistics and Computing | 2016年 / 26卷
关键词
Bayesian inference; Convex analysis; High-dimensional statistics; Markov chain Monte Carlo; Proximal algorithms; Signal processing;
D O I
暂无
中图分类号
学科分类号
摘要
This paper presents a new Metropolis-adjusted Langevin algorithm (MALA) that uses convex analysis to simulate efficiently from high-dimensional densities that are log-concave, a class of probability distributions that is widely used in modern high-dimensional statistics and data analysis. The method is based on a new first-order approximation for Langevin diffusions that exploits log-concavity to construct Markov chains with favourable convergence properties. This approximation is closely related to Moreau–Yoshida regularisations for convex functions and uses proximity mappings instead of gradient mappings to approximate the continuous-time process. The proposed method complements existing MALA methods in two ways. First, the method is shown to have very robust stability properties and to converge geometrically for many target densities for which other MALA are not geometric, or only if the step size is sufficiently small. Second, the method can be applied to high-dimensional target densities that are not continuously differentiable, a class of distributions that is increasingly used in image processing and machine learning and that is beyond the scope of existing MALA and HMC algorithms. To use this method it is necessary to compute or to approximate efficiently the proximity mappings of the logarithm of the target density. For several popular models, including many Bayesian models used in modern signal and image processing and machine learning, this can be achieved with convex optimisation algorithms and with approximations based on proximal splitting techniques, which can be implemented in parallel. The proposed method is demonstrated on two challenging high-dimensional and non-differentiable models related to image resolution enhancement and low-rank matrix estimation that are not well addressed by existing MCMC methodology.
引用
收藏
页码:745 / 760
页数:15
相关论文
共 76 条
[1]  
Afonso M(2011)An augmented Lagrangian approach to the constrained optimization formulation of imaging inverse problems IEEE. Trans. Image Process. 20 681-695
[2]  
Bioucas-Dias J(2012)Fast global convergence of gradient methods for high-dimensional statistical recovery Ann. Stat. 40 2452-2482
[3]  
Figueiredo M(2006)An adaptive version for the Metropolis adjusted Langevin algorithm with a truncated drift Methodol. Comput. Appl. Probab. 8 235-254
[4]  
Agarwal A(2009)NESTA: a fast and accurate first-order method for sparse recovery SIAM J. Imaging Sci. 4 1-39
[5]  
Negahban S(2009)The power of convex relaxation: near-optimal matrix completion IEEE Trans. Inf. Theory 56 2053-2080
[6]  
Wainwright JM(2008)An introduction to compressive sampling IEEE Signal Process. Mag. 25 21-30
[7]  
Atchade Y(2013)Unbiased risk estimates for singular value thresholding and spectral estimators IEEE Trans. Signal Process. 61 4643-4657
[8]  
Becker S(2011)Stability of partially implicit Langevin schemes and their MCMC variants Methodol. Comput. Appl. Probab. 13 835-854
[9]  
Bobin J(2004)An algorithm for total variation minimization and applications J. Math. Imaging Vis. 20 89-97
[10]  
Candès EJ(2013)Computational and statistical tradeoffs via convex relaxation Proc. Natl Acad. Sci. U.S.A. 110 1181-1190