Robust Bayesian model selection for variable clustering with the Gaussian graphical model

被引:2
作者
Andrade, Daniel [1 ,2 ]
Takeda, Akiko [3 ,4 ]
Fukumizu, Kenji [5 ]
机构
[1] SOKENDAI, 10-3 Midoricho, Tachikawa, Tokyo 1908562, Japan
[2] NEC Corp Ltd, Secur Res Labs, 1753 Shimonumabe, Kawasaki, Kanagawa 2118666, Japan
[3] Univ Tokyo, Dept Creat Informat, Bunkyo Ku, 7-3-1 Hongo, Tokyo 1138656, Japan
[4] RIKEN Ctr Adv Intelligence Project, Chuo Ku, 1-4-1 Nihonbashi, Tokyo 1030027, Japan
[5] Inst Stat Math, 10-3 Midoricho, Tachikawa, Tokyo 1908562, Japan
基金
日本学术振兴会;
关键词
Clustering; Gaussian graphical model; Model selection; Variational approximation; INFORMATION CRITERIA; MARGINAL LIKELIHOOD; STOCHASTIC SEARCH;
D O I
10.1007/s11222-019-09879-9
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Variable clustering is important for explanatory analysis. However, only few dedicated methods for variable clustering with the Gaussian graphical model have been proposed. Even more severe, small insignificant partial correlations due to noise can dramatically change the clustering result when evaluating for example with the Bayesian information criteria (BIC). In this work, we try to address this issue by proposing a Bayesian model that accounts for negligible small, but not necessarily zero, partial correlations. Based on our model, we propose to evaluate a variable clustering result using the marginal likelihood. To address the intractable calculation of the marginal likelihood, we propose two solutions: one based on a variational approximation and another based on MCMC. Experiments on simulated data show that the proposed method is similarly accurate as BIC in the no noise setting, but considerably more accurate when there are noisy partial correlations. Furthermore, on real data the proposed method provides clustering results that are intuitively sensible, which is not always the case when using BIC or its extensions.
引用
收藏
页码:351 / 376
页数:26
相关论文
共 36 条
[1]  
Albersts B, 2014, MOL BIOL CELL PROBLE
[2]  
Anderson T., 2004, INTRO MULTIVARIATE S
[3]  
[Anonymous], 2010, NIPS
[4]  
[Anonymous], 1973, 2 INT S INF THEOR
[5]  
[Anonymous], 2012, Adv Neural Inf Process Syst
[6]  
[Anonymous], 2009, P 26 ANN INT C MACHI
[7]   Laplacian eigenmaps for dimensionality reduction and data representation [J].
Belkin, M ;
Niyogi, P .
NEURAL COMPUTATION, 2003, 15 (06) :1373-1396
[8]   Distributed optimization and statistical learning via the alternating direction method of multipliers [J].
Boyd S. ;
Parikh N. ;
Chu E. ;
Peleato B. ;
Eckstein J. .
Foundations and Trends in Machine Learning, 2010, 3 (01) :1-122
[9]  
Brent R.P, 1971, TECHNICAL REPORT
[10]   General methods for monitoring convergence of iterative simulations [J].
Brooks, SP ;
Gelman, A .
JOURNAL OF COMPUTATIONAL AND GRAPHICAL STATISTICS, 1998, 7 (04) :434-455