Scalable Bayesian High-dimensional Local Dependence Learning

被引:1
|
作者
Lee, Kyoungjae [1 ]
Lin, Lizhen [2 ]
机构
[1] Sungkyunkwan Univ, Dept Stat, Seoul, South Korea
[2] Univ Notre Dame, Dept Appl & Computat Math & Stat, Notre Dame, IN USA
来源
BAYESIAN ANALYSIS | 2023年 / 18卷 / 01期
基金
新加坡国家研究基金会;
关键词
selection consistency; optimal posterior convergence rate; varying bandwidth; POSTERIOR CONVERGENCE-RATES; LARGE PRECISION MATRICES; CONSISTENCY; SELECTION; MODELS; LIKELIHOOD; SPARSITY;
D O I
10.1214/21-BA1299
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
In this work, we propose a scalable Bayesian procedure for learning the local dependence structure in a high-dimensional model where the variables possess a natural ordering. The ordering of variables can be indexed by time, the vicinities of spatial locations, and so on, with the natural assumption that variables far apart tend to have weak correlations. Applications of such models abound in a variety of fields such as finance, genome associations analysis and spatial model-ing. We adopt a flexible framework under which each variable is dependent on its neighbors or predecessors, and the neighborhood size can vary for each variable. It is of great interest to reveal this local dependence structure by estimating the covariance or precision matrix while yielding a consistent estimate of the varying neighborhood size for each variable. The existing literature on banded covariance matrix estimation, which assumes a fixed bandwidth cannot be adapted for this general setup. We employ the modified Cholesky decomposition for the precision matrix and design a flexible prior for this model through appropriate priors on the neighborhood sizes and Cholesky factors. The posterior contraction rates of the Cholesky factor are derived which are nearly or exactly minimax optimal, and our procedure leads to consistent estimates of the neighborhood size for all the vari-ables. Another appealing feature of our procedure is its scalability to models with large numbers of variables due to efficient posterior inference without resorting to MCMC algorithms. Numerical comparisons are carried out with competitive methods, and applications are considered for some real datasets.
引用
收藏
页码:25 / 47
页数:23
相关论文
共 50 条
  • [41] High-dimensional posterior consistency of the Bayesian lasso
    Dasgupta, Shibasish
    COMMUNICATIONS IN STATISTICS-THEORY AND METHODS, 2016, 45 (22) : 6700 - 6708
  • [42] Bayesian high-dimensional screening via MCMC
    Shang, Zuofeng
    Li, Ping
    JOURNAL OF STATISTICAL PLANNING AND INFERENCE, 2014, 155 : 54 - 78
  • [43] Frequency-sensitive competitive learning for scalable balanced clustering on high-dimensional hyperspheres
    Banerjee, A
    Ghosh, J
    IEEE TRANSACTIONS ON NEURAL NETWORKS, 2004, 15 (03): : 702 - 719
  • [44] Bayesian Pairwise Comparison of High-Dimensional Images
    Guha, Subharup
    Qiu, Peihua
    JOURNAL OF COMPUTATIONAL AND GRAPHICAL STATISTICS, 2025,
  • [45] Fast and Scalable Learning of Sparse Changes in High-Dimensional Gaussian Graphical Model Structure
    Wang, Beilun
    Sekhon, Arshdeep
    Qi, Yanjun
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 84, 2018, 84
  • [46] SCALABLE PROBABILISTIC MODELING AND MACHINE LEARNING WITH DIMENSIONALITY REDUCTION FOR EXPENSIVE HIGH-DIMENSIONAL PROBLEMS
    Luan, Lele
    Ramachandra, Nesar
    Ravi, Sandipp Krishnan
    Bhaduri, Anindya
    Pandita, Piyush
    Balaprakash, Prasanna
    Anitescu, Mihai
    Sun, Changjie
    Wang, Liping
    PROCEEDINGS OF ASME 2023 INTERNATIONAL DESIGN ENGINEERING TECHNICAL CONFERENCES AND COMPUTERS AND INFORMATION IN ENGINEERING CONFERENCE, IDETC-CIE2023, VOL 2, 2023,
  • [47] A Fast and Scalable Computational Framework for Large-Scale High-Dimensional Bayesian Optimal Experimental Design*
    Wu, Keyi
    Chen, Peng
    Ghattas, Omar
    SIAM-ASA JOURNAL ON UNCERTAINTY QUANTIFICATION, 2023, 11 (01): : 235 - 261
  • [48] μPhos: a scalable and sensitive platform for high-dimensional phosphoproteomics
    Oliinyk, Denys
    Will, Andreas
    Schneidmadel, Felix R.
    Boehme, Maximilian
    Rinke, Jenny
    Hochhaus, Andreas
    Ernst, Thomas
    Hahn, Nina
    Geis, Christian
    Lubeck, Markus
    Raether, Oliver
    Humphrey, Sean J.
    Meier, Florian
    MOLECULAR SYSTEMS BIOLOGY, 2024, 20 (08) : 972 - 995
  • [49] Scalable Collapsed Inference for High-Dimensional Topic Models
    Islam, Rashidul
    Foulds, James
    2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, 2019, : 2836 - 2845
  • [50] Scalable high-dimensional dynamic stochastic economic modeling
    Brumm, Johannes
    Mikushin, Dmitry
    Scheidegger, Simon
    Schenk, Olaf
    JOURNAL OF COMPUTATIONAL SCIENCE, 2015, 11 : 12 - 25