Bayesian Inference Using the Proximal Mapping: Uncertainty Quantification Under Varying Dimensionality

被引:0
作者
Xu, Maoran [1 ]
Zhou, Hua [2 ,3 ]
Hu, Yujie [4 ]
Duan, Leo L. [5 ]
机构
[1] Duke Univ, Dept Stat Sci, Durham, NC 27708 USA
[2] Univ Calif Los Angeles, Dept Biostat, Los Angeles, CA USA
[3] Univ Calif Los Angeles, Dept Computat Med, Los Angeles, CA USA
[4] Univ Florida, Dept Geog, Gainesville, FL USA
[5] Univ Florida, Dept Stat, Gainesville, FL USA
基金
美国国家科学基金会;
关键词
Concentration of Lipschitz functions; Generalized density; Generalized projection; Hausdorff dimension; Nonexpansiveness; VARIABLE SELECTION; LASSO;
D O I
10.1080/01621459.2023.2220170
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
In statistical applications, it is common to encounter parameters supported on a varying or unknown dimensional space. Examples include the fused lasso regression, the matrix recovery under an unknown low rank, etc. Despite the ease of obtaining a point estimate via optimization, it is much more challenging to quantify their uncertainty. In the Bayesian framework, a major difficulty is that if assigning the prior associated with a p-dimensional measure, then there is zero posterior probability on any lower-dimensional subset with dimension d < p. To avoid this caveat, one needs to choose another dimension-selection prior on d, which often involves a highly combinatorial problem. To significantly reduce the modeling burden, we propose a new generative process for the prior: starting from a continuous random variable such as multivariate Gaussian, we transform it into a varying-dimensional space using the proximal mapping. This leads to a large class of new Bayesian models that can directly exploit the popular frequentist regularizations and their algorithms, such as the nuclear norm penalty and the alternating direction method of multipliers, while providing a principled and probabilistic uncertainty estimation. We show that this framework is well justified in the geometric measure theory, and enjoys a convenient posterior computation via the standard Hamiltonian Monte Carlo. We demonstrate its use in the analysis of the dynamic flow network data. for this article are available online.
引用
收藏
页码:1847 / 1858
页数:12
相关论文
共 42 条
[1]  
[Anonymous], 2006, 703 UC BERK DEP STAT
[2]  
Beck A, 2017, MOS-SIAM SER OPTIMIZ, P1, DOI 10.1137/1.9781611974997
[3]   THE FORMAL DEFINITION OF REFERENCE PRIORS [J].
Berger, James O. ;
Bernardo, Jose M. ;
Sun, Dongchu .
ANNALS OF STATISTICS, 2009, 37 (02) :905-938
[4]  
Bertsekas D.P., 1982, CONSTRAINED OPTIMIZA
[5]   Sparse Bayesian infinite factor models [J].
Bhattacharya, A. ;
Dunson, D. B. .
BIOMETRIKA, 2011, 98 (02) :291-306
[6]   SIMULTANEOUS ANALYSIS OF LASSO AND DANTZIG SELECTOR [J].
Bickel, Peter J. ;
Ritov, Ya'acov ;
Tsybakov, Alexandre B. .
ANNALS OF STATISTICS, 2009, 37 (04) :1705-1732
[7]  
Carvalho CM., 2009, ARTIF INTELL, P73
[8]  
Cook RD, 2010, STAT SINICA, V20, P927
[9]   Bayesian curve-fitting with free-knot splines [J].
DiMatteo, I ;
Genovese, CR ;
Kass, RE .
BIOMETRIKA, 2001, 88 (04) :1055-1071
[10]   Bayesian inference on order-constrained parameters in generalized linear models [J].
Dunson, DB .
BIOMETRICS, 2003, 59 (02) :286-295