Gaussian mixture model with feature selection: An embedded approach

被引:30
|
作者
Fu, Yinlin [1 ]
Liu, Xiaonan [1 ]
Sarkar, Suryadipto [1 ]
Wu, Teresa [1 ]
机构
[1] Arizona State Univ, Sch Comp, Informat, Decis Syst Engn, 699 South Mill Ave, Tempe, AZ 85281 USA
关键词
Gaussian Mixture Model (GMM); Expectation Maximization (EM); Feature selection; VARIABLE SELECTION; EM;
D O I
10.1016/j.cie.2020.107000
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Gaussian Mixture Model (GMM) is a popular clustering algorithm due to its neat statistical properties, which enable the "soft" clustering and the determination of the number of clusters. Expectation-Maximization (EM) is usually applied to estimate the GMM parameters. While promising, the inclusion of features that are not contributing to clustering may confuse the model and increase computational cost. Recognizing the issue, in this paper, we propose a new algorithm, termed Expectation Selection Maximization (ESM), by adding a feature selection step (5). Specifically, we introduce a relevancy index (RI), a metric indicating the probability of assigning a data point to a specific clustering group. The RI index reveals the contribution of the feature to the clustering process thus can assist the feature selection. We conduct theoretical analysis to justify the use of RI for feature selection. Also, to demonstrate the efficacy of the proposed ESM, two synthetic datasets, four benchmark datasets, and an Alzheimer's Disease dataset are studied.
引用
收藏
页数:9
相关论文
共 50 条
  • [31] Feature Guided Biased Gaussian Mixture Model for image matching
    Sun, Kun
    Li, Peiran
    Tao, Wenbing
    Tang, Yuanyan
    INFORMATION SCIENCES, 2015, 295 : 323 - 336
  • [32] Feature Analysis on English word difficulty by Gaussian Mixture Model
    Yang, Hua
    Eum, Suyong
    2018 INTERNATIONAL CONFERENCE ON INFORMATION AND COMMUNICATION TECHNOLOGY CONVERGENCE (ICTC), 2018, : 191 - 194
  • [33] Asymmetric Mixture Model With Simultaneous Feature Selection and Model Detection
    Thanh Minh Nguyen
    Wu, Q. M. Jonathan
    Zhang, Hui
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2015, 26 (02) : 400 - 408
  • [34] A survey of feature selection methods for Gaussian mixture models and hidden Markov models
    Stephen Adams
    Peter A. Beling
    Artificial Intelligence Review, 2019, 52 : 1739 - 1779
  • [35] Fast Forward Feature Selection of Hyperspectral Images for Classification With Gaussian Mixture Models
    Fauvel, Mathieu
    Dechesne, Clement
    Zullo, Anthony
    Ferraty, Frederic
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2015, 8 (06) : 2824 - 2831
  • [36] A survey of feature selection methods for Gaussian mixture models and hidden Markov models
    Adams, Stephen
    Beling, Peter A.
    ARTIFICIAL INTELLIGENCE REVIEW, 2019, 52 (03) : 1739 - 1779
  • [37] Feature selection for pattern classification with Gaussian mixture models: A new objective criterion
    Krishnan, S
    Samudravijaya, K
    Rao, PVS
    PATTERN RECOGNITION LETTERS, 1996, 17 (08) : 803 - 809
  • [38] A countably infinite mixture model for clustering and feature selection
    Nizar Bouguila
    Djemel Ziou
    Knowledge and Information Systems, 2012, 33 : 351 - 370
  • [39] Model Selection Criterion for Multivariate Bounded Asymmetric Gaussian Mixture Model
    Xian, Zixiang
    Azam, Muhammad
    Amayri, Manar
    Bouguila, Nizar
    29TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO 2021), 2021, : 1436 - 1440
  • [40] A countably infinite mixture model for clustering and feature selection
    Bouguila, Nizar
    Ziou, Djemel
    KNOWLEDGE AND INFORMATION SYSTEMS, 2012, 33 (02) : 351 - 370