Scalable Bayesian High-dimensional Local Dependence Learning

被引:1
|
作者
Lee, Kyoungjae [1 ]
Lin, Lizhen [2 ]
机构
[1] Sungkyunkwan Univ, Dept Stat, Seoul, South Korea
[2] Univ Notre Dame, Dept Appl & Computat Math & Stat, Notre Dame, IN USA
来源
BAYESIAN ANALYSIS | 2023年 / 18卷 / 01期
基金
新加坡国家研究基金会;
关键词
selection consistency; optimal posterior convergence rate; varying bandwidth; POSTERIOR CONVERGENCE-RATES; LARGE PRECISION MATRICES; CONSISTENCY; SELECTION; MODELS; LIKELIHOOD; SPARSITY;
D O I
10.1214/21-BA1299
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
In this work, we propose a scalable Bayesian procedure for learning the local dependence structure in a high-dimensional model where the variables possess a natural ordering. The ordering of variables can be indexed by time, the vicinities of spatial locations, and so on, with the natural assumption that variables far apart tend to have weak correlations. Applications of such models abound in a variety of fields such as finance, genome associations analysis and spatial model-ing. We adopt a flexible framework under which each variable is dependent on its neighbors or predecessors, and the neighborhood size can vary for each variable. It is of great interest to reveal this local dependence structure by estimating the covariance or precision matrix while yielding a consistent estimate of the varying neighborhood size for each variable. The existing literature on banded covariance matrix estimation, which assumes a fixed bandwidth cannot be adapted for this general setup. We employ the modified Cholesky decomposition for the precision matrix and design a flexible prior for this model through appropriate priors on the neighborhood sizes and Cholesky factors. The posterior contraction rates of the Cholesky factor are derived which are nearly or exactly minimax optimal, and our procedure leads to consistent estimates of the neighborhood size for all the vari-ables. Another appealing feature of our procedure is its scalability to models with large numbers of variables due to efficient posterior inference without resorting to MCMC algorithms. Numerical comparisons are carried out with competitive methods, and applications are considered for some real datasets.
引用
收藏
页码:25 / 47
页数:23
相关论文
共 50 条
  • [31] COMPLEXITY ANALYSIS OF BAYESIAN LEARNING OF HIGH-DIMENSIONAL DAG MODELS AND THEIR EQUIVALENCE CLASSES
    Zhou, Quan
    Chang, Hyunwoong
    ANNALS OF STATISTICS, 2023, 51 (03): : 1058 - 1085
  • [32] Learning Moral Graphs in Construction of High-Dimensional Bayesian Networks for Mixed Data
    Xu, Suwa
    Jia, Bochao
    Liang, Faming
    NEURAL COMPUTATION, 2019, 31 (06) : 1183 - 1214
  • [33] Learning Discriminative Bayesian Networks from High-Dimensional Continuous Neuroimaging Data
    Zhou, Luping
    Wang, Lei
    Liu, Lingqiao
    Ogunbona, Philip
    Shen, Dinggang
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2016, 38 (11) : 2269 - 2283
  • [34] Ensemble-Based Local Learning for High-Dimensional Data Regression
    Raytchev, B.
    Katamoto, Y.
    Koujiba, M.
    Tamaki, T.
    Kaneda, K.
    2016 23RD INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2016, : 2640 - 2645
  • [35] High-Dimensional Bayesian Network Classification with Network Global-Local Shrinkage Priors
    Guha, Sharmistha
    Rodriguez, Abel
    BAYESIAN ANALYSIS, 2023, 18 (04): : 1131 - 1160
  • [36] An asymptotic viewpoint on high-dimensional Bayesian testing
    Spitzner, Dan J.
    BAYESIAN ANALYSIS, 2008, 3 (01): : 121 - 160
  • [37] Bayesian Model Selection in High-Dimensional Settings
    Johnson, Valen E.
    Rossell, David
    JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2012, 107 (498) : 649 - 660
  • [38] Bayesian inference for high-dimensional decomposable graphs
    Lee, Kyoungjae
    Cao, Xuan
    ELECTRONIC JOURNAL OF STATISTICS, 2021, 15 (01): : 1549 - 1582
  • [39] A nonparametric Bayesian technique for high-dimensional regression
    Guha, Subharup
    Baladandayuthapani, Veerabhadran
    ELECTRONIC JOURNAL OF STATISTICS, 2016, 10 (02): : 3374 - 3424
  • [40] Learning high-dimensional data
    Verleysen, M
    LIMITATIONS AND FUTURE TRENDS IN NEURAL COMPUTATION, 2003, 186 : 141 - 162