A Variational Approximations-DIC Rubric for Parameter Estimation and Mixture Model Selection Within a Family Setting

被引:6
作者
Subedi, Sanjeena [1 ]
McNicholas, Paul D. [2 ]
机构
[1] SUNY Binghamton, Dept Math Sci, 4400 Vestal Pkwy East, Binghamton, NY 13902 USA
[2] McMaster Univ, Dept Math & Stat, 1280 Main St W, Hamilton, ON L8S 4K1, Canada
关键词
BIC; Clustering; DIC; EM algorithm; GPCM; Mixture models; Model-based clustering; Variational approximations; Variational Bayes; VB-DIC; BAYESIAN-ANALYSIS; UNKNOWN NUMBER; MULTIVARIATE; COMPONENTS; INFERENCE; DIMENSION; CLASSIFICATION; LIKELIHOOD;
D O I
10.1007/s00357-019-09351-3
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
Mixture model-based clustering has become an increasingly popular data analysis technique since its introduction over fifty years ago, and is now commonly utilized within a family setting. Families of mixture models arise when the component parameters, usually the component covariance (or scale) matrices, are decomposed and a number of constraints are imposed. Within the family setting, model selection involves choosing the member of the family, i.e., the appropriate covariance structure, in addition to the number of mixture components. To date, the Bayesian information criterion (BIC) has proved most effective for model selection, and the expectation-maximization (EM) algorithm is usually used for parameter estimation. In fact, this EM-BIC rubric has virtually monopolized the literature on families of mixture models. Deviating from this rubric, variational Bayes approximations are developed for parameter estimation and the deviance information criteria (DIC) for model selection. The variational Bayes approach provides an alternate framework for parameter estimation by constructing a tight lower bound on the complex marginal likelihood and maximizing this lower bound by minimizing the associated Kullback-Leibler divergence. The framework introduced, which we refer to as VB-DIC, is applied to the most commonly used family of Gaussian mixture models, and real and simulated data are used to compared with the EM-BIC rubric.
引用
收藏
页码:89 / 108
页数:20
相关论文
共 84 条
[1]  
Aitken A.C., 1926, Proc. R. Soc. Edinb., V45, P14
[2]   STOCHASTIC THEORY OF MINIMAL REALIZATION [J].
AKAIKE, H .
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 1974, AC19 (06) :667-674
[3]  
[Anonymous], 1965, USDA TECHNICAL B
[4]   MODEL-BASED GAUSSIAN AND NON-GAUSSIAN CLUSTERING [J].
BANFIELD, JD ;
RAFTERY, AE .
BIOMETRICS, 1993, 49 (03) :803-821
[5]   Inference in model-based cluster analysis [J].
Bensmail, H ;
Celeux, G ;
Raftery, AE ;
Robert, CP .
STATISTICS AND COMPUTING, 1997, 7 (01) :1-10
[6]   Assessing a mixture model for clustering with the integrated completed likelihood [J].
Biernacki, C ;
Celeux, G ;
Govaert, G .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2000, 22 (07) :719-725
[7]   Unifying data units and models in (co-)clustering [J].
Biernacki, Christophe ;
Lourme, Alexandre .
ADVANCES IN DATA ANALYSIS AND CLASSIFICATION, 2019, 13 (01) :7-31
[8]   ANTIPODALLY SYMMETRIC DISTRIBUTION ON SPHERE [J].
BINGHAM, C .
ANNALS OF STATISTICS, 1974, 2 (06) :1201-1225
[9]   Variational Inference: A Review for Statisticians [J].
Blei, David M. ;
Kucukelbir, Alp ;
McAuliffe, Jon D. .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2017, 112 (518) :859-877
[10]  
Bock HansH., 1998, DATA SCI CLASSIFICAT, P3