Hierarchical Bayes based Adaptive Sparsity in Gaussian Mixture Model

被引:0
作者
Wang, Binghui [1 ]
Lin, Chuang [1 ,2 ]
Fan, Xin [1 ]
Jiang, Ning [2 ]
Farina, Dario [2 ]
机构
[1] Dalian Univ Technol, Sch Software, Dalian, Peoples R China
[2] Univ Gottingen, Univ Med Ctr Goettingen, Dept Neurorehabil Engn, D-37073 Gottingen, Germany
基金
欧洲研究理事会;
关键词
High-dimensional parameter estimation; Hierarchical Bayes; Adaptive sparsity; GMM; COVARIANCE-MATRIX ESTIMATION; CONVERGENCE; SELECTION; RATES;
D O I
10.1016/j.patrec.2014.07.008
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Gaussian Mixture Model (GMM) has been widely used in statistics for its great flexibility. However, parameter estimation for GMM with high dimensionality is a challenge because of the large number of parameters and the lack of observation data. In this paper, we propose an effective method named hierarchical Bayes based Adaptive Sparsity in Gaussian Mixture Model (ASGMM) to estimate the parameters in a GMM by incorporating a two-layer hierarchical Bayes based adaptive sparsity prior. The prior we impose on the precision matrices can encourage sparsity and hence reduce the dimensionality of the parameters to be estimated. In contrast to the l(1)-norm penalty or Laplace prior, our approach does not involve any hyperparameters that must be tuned, and the sparsity adapts to the observation data. The proposed method is achieved by three steps: first, we formulate an adaptive hierarchical Bayes model of the precision matrices in the GMM with a Jeffrey's noninformative hyperprior, which expresses scale-invariance and, more importantly, is hyperparameter-free and unbiased. Second, we perform a Cholesky decomposition on the precision matrices to impose the positive definite property. Finally, we exploit the expectation maximization (EM) algorithm to obtain the final estimated parameters in the GMM. Experimental results on synthetic and real-world datasets demonstrate that ASGMM cannot only adapt the sparsity of high-dimensional data with small estimated error, but also achieve better clustering performance comparing with several classical methods. (C) 2014 Elsevier B.V. All rights reserved.
引用
收藏
页码:238 / 247
页数:10
相关论文
共 28 条
[21]   High-dimensional graphs and variable selection with the Lasso [J].
Meinshausen, Nicolai ;
Buehlmann, Peter .
ANNALS OF STATISTICS, 2006, 34 (03) :1436-1462
[22]   High-dimensional covariance estimation by minimizing l1-penalized log-determinant divergence [J].
Ravikumar, Pradeep ;
Wainwright, Martin J. ;
Raskutti, Garvesh ;
Yu, Bin .
ELECTRONIC JOURNAL OF STATISTICS, 2011, 5 :935-980
[23]   Sparse permutation invariant covariance estimation [J].
Rothman, Adam J. ;
Bickel, Peter J. ;
Levina, Elizaveta ;
Zhu, Ji .
ELECTRONIC JOURNAL OF STATISTICS, 2008, 2 :494-515
[24]   Regularized Parameter Estimation in High-Dimensional Gaussian Mixture Models [J].
Ruan, Lingyan ;
Yuan, Ming ;
Zou, Hui .
NEURAL COMPUTATION, 2011, 23 (06) :1605-1622
[25]   ON SCALE MIXTURES OF NORMAL-DISTRIBUTIONS [J].
WEST, M .
BIOMETRIKA, 1987, 74 (03) :646-648
[26]  
Wong Eleanor., 2013, International Conference on Machine Learning (ICML), P311
[27]   Model selection and estimation in the Gaussian graphical model [J].
Yuan, Ming ;
Lin, Yi .
BIOMETRIKA, 2007, 94 (01) :19-35
[28]  
Yuan M, 2010, J MACH LEARN RES, V11, P2261