A countably infinite mixture model for clustering and feature selection

被引:37
作者
Bouguila, Nizar [1 ]
Ziou, Djemel [2 ]
机构
[1] Concordia Univ, CIISE, Montreal, PQ H3G 1T7, Canada
[2] Univ Sherbrooke, Sherbrooke, PQ J1K 2R1, Canada
关键词
Non-parametric Bayesian methods; Dirichlet process; Clustering; Feature selection; Mixture models; Generalized Dirichlet; MCMC; Categorization; GENERALIZED DIRICHLET MIXTURE; UNSUPERVISED SELECTION; SCENE;
D O I
10.1007/s10115-011-0467-4
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Mixture modeling is one of the most useful tools in machine learning and data mining applications. An important challenge when applying finite mixture models is the selection of the number of clusters which best describes the data. Recent developments have shown that this problem can be handled by the application of non-parametric Bayesian techniques to mixture modeling. Another important crucial preprocessing step to mixture learning is the selection of the most relevant features. The main approach in this paper, to tackle these problems, consists on storing the knowledge in a generalized Dirichlet mixture model by applying non-parametric Bayesian estimation and inference techniques. Specifically, we extend finite generalized Dirichlet mixture models to the infinite case in which the number of components and relevant features do not need to be known a priori. This extension provides a natural representation of uncertainty regarding the challenging problem of model selection. We propose a Markov Chain Monte Carlo algorithm to learn the resulted infinite mixture. Through applications involving text and image categorization, we show that infinite mixture models offer a more powerful and robust performance than classic finite mixtures for both clustering and feature selection.
引用
收藏
页码:351 / 370
页数:20
相关论文
共 38 条
[1]  
[Anonymous], 2003, Bayesian Data Analysis
[2]  
[Anonymous], 2003, BAYESIAN NONPARAMETR
[3]  
[Anonymous], 1983, RECNT ADV STAT
[4]  
Bartolini I, 2009, KNOWL INF SYST IN PR
[5]   Practical Bayesian estimation of a finite beta mixture through gibbs sampling and its applications [J].
Bouguila, N ;
Ziou, D ;
Monga, E .
STATISTICS AND COMPUTING, 2006, 16 (02) :215-225
[6]   A powreful finite mixture model based on the generalized Dirichlet distribution: Unsupervised learning and applications [J].
Bouguila, N ;
Ziou, D .
PROCEEDINGS OF THE 17TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION, VOL 1, 2004, :280-283
[7]  
Bouguila N, 2004, 2004 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOL V, PROCEEDINGS, P521
[8]  
Bouguila N, 2003, LECT NOTES ARTIF INT, V2734, P172
[9]   Clustering of count data using generalized Dirichlet multinomial distributions [J].
Bouguila, Nizar .
IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2008, 20 (04) :462-474
[10]  
Bouguila N, 2007, IEEE T PATTERN ANAL, V29, P1716, DOI [10.1109/TPAMI.2007.1095, 10.1109/TPAMl.2007.1095]