Scalable Bayesian High-dimensional Local Dependence Learning

被引:1
|
作者
Lee, Kyoungjae [1 ]
Lin, Lizhen [2 ]
机构
[1] Sungkyunkwan Univ, Dept Stat, Seoul, South Korea
[2] Univ Notre Dame, Dept Appl & Computat Math & Stat, Notre Dame, IN USA
来源
BAYESIAN ANALYSIS | 2023年 / 18卷 / 01期
基金
新加坡国家研究基金会;
关键词
selection consistency; optimal posterior convergence rate; varying bandwidth; POSTERIOR CONVERGENCE-RATES; LARGE PRECISION MATRICES; CONSISTENCY; SELECTION; MODELS; LIKELIHOOD; SPARSITY;
D O I
10.1214/21-BA1299
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
In this work, we propose a scalable Bayesian procedure for learning the local dependence structure in a high-dimensional model where the variables possess a natural ordering. The ordering of variables can be indexed by time, the vicinities of spatial locations, and so on, with the natural assumption that variables far apart tend to have weak correlations. Applications of such models abound in a variety of fields such as finance, genome associations analysis and spatial model-ing. We adopt a flexible framework under which each variable is dependent on its neighbors or predecessors, and the neighborhood size can vary for each variable. It is of great interest to reveal this local dependence structure by estimating the covariance or precision matrix while yielding a consistent estimate of the varying neighborhood size for each variable. The existing literature on banded covariance matrix estimation, which assumes a fixed bandwidth cannot be adapted for this general setup. We employ the modified Cholesky decomposition for the precision matrix and design a flexible prior for this model through appropriate priors on the neighborhood sizes and Cholesky factors. The posterior contraction rates of the Cholesky factor are derived which are nearly or exactly minimax optimal, and our procedure leads to consistent estimates of the neighborhood size for all the vari-ables. Another appealing feature of our procedure is its scalability to models with large numbers of variables due to efficient posterior inference without resorting to MCMC algorithms. Numerical comparisons are carried out with competitive methods, and applications are considered for some real datasets.
引用
收藏
页码:25 / 47
页数:23
相关论文
共 50 条
  • [1] Local Bayesian Network Structure Learning for High-Dimensional Data
    Wang, Yangyang
    Gao, Xiaoguang
    Sun, Pengzhan
    Ru, Xinxin
    Wang, Jihan
    2024 9TH INTERNATIONAL CONFERENCE ON CONTROL AND ROBOTICS ENGINEERING, ICCRE 2024, 2024, : 259 - 263
  • [2] Local Bayesian network structure learning for high-dimensional data
    Wang, Yangyang
    Gao, Xiaoguang
    Ru, Xinxin
    Xi Tong Gong Cheng Yu Dian Zi Ji Shu/Systems Engineering and Electronics, 2024, 46 (08): : 2676 - 2685
  • [3] A scalable approximate Bayesian inference for high-dimensional Gaussian processes
    Fradi, Anis
    Samir, Chafik
    Bachoc, Francois
    COMMUNICATIONS IN STATISTICS-THEORY AND METHODS, 2022, 51 (17) : 5937 - 5956
  • [4] Scalable Bayesian variable selection for structured high-dimensional data
    Chang, Changgee
    Kundu, Suprateek
    Long, Qi
    BIOMETRICS, 2018, 74 (04) : 1372 - 1382
  • [5] MixDir: Scalable Bayesian Clustering for High-Dimensional Categorical Data
    Ahlmann-Eltze, Constantin
    Yau, Christopher
    2018 IEEE 5TH INTERNATIONAL CONFERENCE ON DATA SCIENCE AND ADVANCED ANALYTICS (DSAA), 2018, : 526 - 539
  • [6] Scalable collaborative targeted learning for high-dimensional data
    Ju, Cheng
    Gruber, Susan
    Lendle, Samuel D.
    Chambaz, Antoine
    Franklin, Jessica M.
    Wyss, Richard
    Schneeweiss, Sebastian
    van der Laan, Mark J.
    STATISTICAL METHODS IN MEDICAL RESEARCH, 2019, 28 (02) : 532 - 554
  • [7] Scalable Algorithms for Learning High-Dimensional Linear Mixed Models
    Tan, Zilong
    Roche, Kimberly
    Zhou, Xiang
    Mukherjee, Sayan
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2018, : 259 - 268
  • [8] Conditional dependence learning with high-dimensional conditioning variables
    Bi, Jianxin
    Feng, Xingdong
    Liu, Jingyuan
    SCIENCE CHINA-MATHEMATICS, 2025,
  • [9] Uncertainty Quantification for Modern High-Dimensional Regression via Scalable Bayesian Methods
    Rajaratnam, Bala
    Sparks, Doug
    Khare, Kshitij
    Zhang, Liyuan
    JOURNAL OF COMPUTATIONAL AND GRAPHICAL STATISTICS, 2019, 28 (01) : 174 - 184
  • [10] Scalable spatio-temporal Bayesian analysis of high-dimensional electroencephalography data
    Mohammed, Shariq
    Dey, Dipak K.
    CANADIAN JOURNAL OF STATISTICS-REVUE CANADIENNE DE STATISTIQUE, 2021, 49 (01): : 107 - 128