Nearest neighbor estimate of conditional mutual information in feature selection

被引:32
|
作者
Tsimpiris, Alkiviadis [1 ]
Vlachos, Ioannis [2 ]
Kugiumtzis, Dimitris [1 ]
机构
[1] Aristotle Univ Thessaloniki, Fac Engn, Thessaloniki, Greece
[2] Arizona State Univ, Ira Fulton Sch Engn, Sch Biol & Hlth Syst Engn, Tempe, AZ USA
关键词
Feature selection; Conditional mutual information; Nearest neighbor estimate; mRMR; MaxiMin; Classification; INPUT FEATURE-SELECTION; TIME-SERIES; CLASSIFICATION;
D O I
10.1016/j.eswa.2012.05.014
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Mutual information (MI) is used in feature selection to evaluate two key-properties of optimal features, the relevance of a feature to the class variable and the redundancy of similar features. Conditional mutual information (CMI), i.e., MI of the candidate feature to the class variable conditioning on the features already selected, is a natural extension of MI but not so far applied due to estimation complications for high dimensional distributions. We propose the nearest neighbor estimate of CMI, appropriate for high-dimensional variables, and build an iterative scheme for sequential feature selection with a termination criterion, called CMINN. We show that CMINN is equivalent to feature selection MI filters, such as mRMR and MaxiMin, in the presence of solely single feature effects, and more appropriate for combined feature effects. We compare CMINN to mRMR and MaxiMin on simulated datasets involving combined effects and confirm the superiority of CMINN in selecting the correct features (indicated also by the termination criterion) and giving best classification accuracy. The application to ten benchmark databases shows that CMINN obtains the same or higher classification accuracy compared to mRMR and MaxiMin at a smaller cardinality of the selected feature subset. (C) 2012 Elsevier Ltd. All rights reserved.
引用
收藏
页码:12697 / 12708
页数:12
相关论文
共 50 条
  • [31] Breaking boundaries: Low-precision conditional mutual information for efficient feature selection
    Moran-Fernandez, Laura
    Blanco-Mallo, Eva
    Sechidis, Konstantinos
    Bolon-Canedo, Veronica
    PATTERN RECOGNITION, 2025, 162
  • [32] On the benefit of feature selection and ensemble feature selection for fuzzy k-nearest neighbor classification
    Lohrmann, Christoph
    Lohrmann, Alena
    Kumbure, Mahinda Mailagaha
    APPLIED SOFT COMPUTING, 2025, 171
  • [33] Supervised feature selection by clustering using conditional mutual information-based distances
    Martinez Sotoca, Jose
    Pla, Filiberto
    PATTERN RECOGNITION, 2010, 43 (06) : 2068 - 2081
  • [34] Efficient and Intelligent Feature Selection via Maximum Conditional Mutual Information for Microarray Data
    Zhang, Jiangnan
    Li, Shaojing
    Yang, Huaichuan
    Jiang, Jingtao
    Shi, Hongtao
    APPLIED SCIENCES-BASEL, 2024, 14 (13):
  • [35] Geometric k-nearest neighbor estimation of entropy and mutual information
    Lord, Warren M.
    Sun, Jie
    Bollt, Erik M.
    CHAOS, 2018, 28 (03)
  • [36] Nearest Neighbor Classification Method Based on The Mutual Information Distance Measure
    Gao, YunLong
    Yan, Peng
    Pan, JinYan
    2014 11TH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION (WCICA), 2014, : 3246 - 3250
  • [37] A new general nearest neighbor classification based on the mutual neighborhood information
    Pan, Zhibin
    Wang, Yidi
    Ku, Weiping
    KNOWLEDGE-BASED SYSTEMS, 2017, 121 : 142 - 152
  • [38] Quadratic Mutual Information Feature Selection
    Sluga, Davor
    Lotric, Uros
    ENTROPY, 2017, 19 (04)
  • [39] Weighted Mutual Information for Feature Selection
    Schaffernicht, Erik
    Gross, Horst-Michael
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2011, PT II, 2011, 6792 : 181 - 188
  • [40] Mutual Information Criteria for Feature Selection
    Zhang, Zhihong
    Hancock, Edwin R.
    SIMILARITY-BASED PATTERN RECOGNITION: FIRST INTERNATIONAL WORKSHOP, SIMBAD 2011, 2011, 7005 : 235 - 249