A Variational Approximations-DIC Rubric for Parameter Estimation and Mixture Model Selection Within a Family Setting

被引:6
作者
Subedi, Sanjeena [1 ]
McNicholas, Paul D. [2 ]
机构
[1] SUNY Binghamton, Dept Math Sci, 4400 Vestal Pkwy East, Binghamton, NY 13902 USA
[2] McMaster Univ, Dept Math & Stat, 1280 Main St W, Hamilton, ON L8S 4K1, Canada
关键词
BIC; Clustering; DIC; EM algorithm; GPCM; Mixture models; Model-based clustering; Variational approximations; Variational Bayes; VB-DIC; BAYESIAN-ANALYSIS; UNKNOWN NUMBER; MULTIVARIATE; COMPONENTS; INFERENCE; DIMENSION; CLASSIFICATION; LIKELIHOOD;
D O I
10.1007/s00357-019-09351-3
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
Mixture model-based clustering has become an increasingly popular data analysis technique since its introduction over fifty years ago, and is now commonly utilized within a family setting. Families of mixture models arise when the component parameters, usually the component covariance (or scale) matrices, are decomposed and a number of constraints are imposed. Within the family setting, model selection involves choosing the member of the family, i.e., the appropriate covariance structure, in addition to the number of mixture components. To date, the Bayesian information criterion (BIC) has proved most effective for model selection, and the expectation-maximization (EM) algorithm is usually used for parameter estimation. In fact, this EM-BIC rubric has virtually monopolized the literature on families of mixture models. Deviating from this rubric, variational Bayes approximations are developed for parameter estimation and the deviance information criteria (DIC) for model selection. The variational Bayes approach provides an alternate framework for parameter estimation by constructing a tight lower bound on the complex marginal likelihood and maximizing this lower bound by minimizing the associated Kullback-Leibler divergence. The framework introduced, which we refer to as VB-DIC, is applied to the most commonly used family of Gaussian mixture models, and real and simulated data are used to compared with the EM-BIC rubric.
引用
收藏
页码:89 / 108
页数:20
相关论文
共 84 条
[71]  
Scrucca L, 2016, R J, V8, P289
[72]   Bayesian measures of model complexity and fit [J].
Spiegelhalter, DJ ;
Best, NG ;
Carlin, BR ;
van der Linde, A .
JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-STATISTICAL METHODOLOGY, 2002, 64 :583-616
[73]   Bayesian analysis of mixture models with an unknown number of components - An alternative to reversible jump methods [J].
Stephens, M .
ANNALS OF STATISTICS, 2000, 28 (01) :40-74
[74]  
STEPHENS M, 1997, THESIS U OXFORD
[75]  
Subedi S, 2015, STAT METHOD APPL-GER, V24, P623, DOI 10.1007/s10260-015-0298-7
[76]   Variational Bayes approximations for clustering via mixtures of normal inverse Gaussian distributions [J].
Subedi, Sanjeena ;
McNicholas, Paul D. .
ADVANCES IN DATA ANALYSIS AND CLASSIFICATION, 2014, 8 (02) :167-193
[77]   A Mixture of Coalesced Generalized Hyperbolic Distributions [J].
Tortora, Cristina ;
Franczak, Brian C. ;
Browne, Ryan P. ;
McNicholas, Paul D. .
JOURNAL OF CLASSIFICATION, 2019, 36 (01) :26-57
[78]   Bayesian model search for mixture models based on optimizing variational bounds [J].
Ueda, N ;
Ghahramani, Z .
NEURAL NETWORKS, 2002, 15 (10) :1223-1241
[79]  
Venables WN., 2002, Modern applied statistics with S, V4th, DOI DOI 10.1007/978-0-387-21706-2
[80]   Finite mixtures of matrix normal distributions for classifying three-way data [J].
Viroli, Cinzia .
STATISTICS AND COMPUTING, 2011, 21 (04) :511-522