TRUST-TECH-based Expectation Maximization for learning finite mixture models

被引:41
作者
Reddy, Chandan K. [1 ]
Chiang, Hsiao-Dong [2 ]
Rajaratnam, Bala [3 ]
机构
[1] Wayne State Univ, Dept Comp Sci, Detroit, MI 48202 USA
[2] Cornell Univ, Dept Elect & Comp Engn, Ithaca, NY 14853 USA
[3] Stanford Univ, Dept Stat, Stanford, CA 94305 USA
关键词
Expectation maximization; unsupervised learning; finite mixture models; dynamical systems; stability regions; model-based clustering;
D O I
10.1109/TPAMI.2007.70775
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The Expectation Maximization (EM) algorithm is widely used for learning finite mixture models despite its greedy nature. Most popular model-based clustering techniques might yield poor clusters if the parameters are not initialized properly. To reduce the sensitivity of initial points, a novel algorithm for learning mixture models from multivariate data is introduced in this paper. The proposed algorithm takes advantage of TRUST-TECH (TRansformation Under STability-reTaining Equilibria CHaracterization) to compute neighborhood local maxima on the likelihood surface using stability regions. Basically, our method coalesces the advantages of the traditional EM with that of the dynamic and geometric characteristics of the stability regions of the corresponding nonlinear dynamical system of the log-likelihood function. Two phases, namely, the EM phase and the stability region phase, are repeated alternatively in the parameter space to achieve local maxima with improved likelihood values. The EM phase obtains the local maximum of the likelihood function and the stability region phase helps to escape out of the local maximum by moving toward the neighboring stability regions. Though applied to Gaussian mixtures in this paper, our technique can be easily generalized to any other parametric finite mixture model. The algorithm has been tested on both synthetic and real data sets and the improvements in the performance compared to other approaches are demonstrated. The robustness with respect to initialization is also illustrated experimentally.
引用
收藏
页码:1146 / 1157
页数:12
相关论文
共 41 条
[1]  
[Anonymous], UCI REPOSITORY MACHI
[2]  
[Anonymous], A gentle tutorial on the EM algorithm and its application to parameter estimation for Gaussian mixture and hidden Markov models
[3]   MODEL-BASED GAUSSIAN AND NON-GAUSSIAN CLUSTERING [J].
BANFIELD, JD ;
RAFTERY, AE .
BIOMETRICS, 1993, 49 (03) :803-821
[4]   A MAXIMIZATION TECHNIQUE OCCURRING IN STATISTICAL ANALYSIS OF PROBABILISTIC FUNCTIONS OF MARKOV CHAINS [J].
BAUM, LE ;
PETRIE, T ;
SOULES, G ;
WEISS, N .
ANNALS OF MATHEMATICAL STATISTICS, 1970, 41 (01) :164-&
[5]   Blobworld: Image segmentation using expectation-maximization and its application to image querying [J].
Carson, C ;
Belongie, S ;
Greenspan, H ;
Malik, J .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2002, 24 (08) :1026-1038
[6]   A systematic search method for obtaining multiple local optimal solutions of nonlinear programming problems [J].
Chiang, HD ;
Chu, CC .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 1996, 43 (02) :99-109
[7]  
DEMSPTER AP, 1977, J ROYAL STAT SOC B, V39, P38
[8]  
Elidan G, 2002, EIGHTEENTH NATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE (AAAI-02)/FOURTEENTH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE (IAAI-02), PROCEEDINGS, P132
[9]   Unsupervised learning of finite mixture models [J].
Figueiredo, MAT ;
Jain, AK .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2002, 24 (03) :381-396
[10]  
Ghahramani Z., 1996, The EM Algorithm for Mixtures of Factor Analyzers