Nearest neighbor estimate of conditional mutual information in feature selection

被引:32
|
作者
Tsimpiris, Alkiviadis [1 ]
Vlachos, Ioannis [2 ]
Kugiumtzis, Dimitris [1 ]
机构
[1] Aristotle Univ Thessaloniki, Fac Engn, Thessaloniki, Greece
[2] Arizona State Univ, Ira Fulton Sch Engn, Sch Biol & Hlth Syst Engn, Tempe, AZ USA
关键词
Feature selection; Conditional mutual information; Nearest neighbor estimate; mRMR; MaxiMin; Classification; INPUT FEATURE-SELECTION; TIME-SERIES; CLASSIFICATION;
D O I
10.1016/j.eswa.2012.05.014
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Mutual information (MI) is used in feature selection to evaluate two key-properties of optimal features, the relevance of a feature to the class variable and the redundancy of similar features. Conditional mutual information (CMI), i.e., MI of the candidate feature to the class variable conditioning on the features already selected, is a natural extension of MI but not so far applied due to estimation complications for high dimensional distributions. We propose the nearest neighbor estimate of CMI, appropriate for high-dimensional variables, and build an iterative scheme for sequential feature selection with a termination criterion, called CMINN. We show that CMINN is equivalent to feature selection MI filters, such as mRMR and MaxiMin, in the presence of solely single feature effects, and more appropriate for combined feature effects. We compare CMINN to mRMR and MaxiMin on simulated datasets involving combined effects and confirm the superiority of CMINN in selecting the correct features (indicated also by the termination criterion) and giving best classification accuracy. The application to ten benchmark databases shows that CMINN obtains the same or higher classification accuracy compared to mRMR and MaxiMin at a smaller cardinality of the selected feature subset. (C) 2012 Elsevier Ltd. All rights reserved.
引用
收藏
页码:12697 / 12708
页数:12
相关论文
共 50 条
  • [1] Conditional independence testing based on a nearest-neighbor estimator of conditional mutual information
    Runge, Jakob
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 84, 2018, 84
  • [2] FEATURE SELECTION WITH WEIGHTED CONDITIONAL MUTUAL INFORMATION
    Celik, Ceyhun
    Bilge, Hasan Sakir
    JOURNAL OF THE FACULTY OF ENGINEERING AND ARCHITECTURE OF GAZI UNIVERSITY, 2015, 30 (04): : 585 - 596
  • [3] Conditional Mutual Information based Feature Selection
    Cheng, Hongrong
    Qin, Zhiguang
    Qian, Weizhong
    Liu, Wei
    KAM: 2008 INTERNATIONAL SYMPOSIUM ON KNOWLEDGE ACQUISITION AND MODELING, PROCEEDINGS, 2008, : 103 - 107
  • [4] Feature Selection with Conditional Mutual Information Considering Feature Interaction
    Liang, Jun
    Hou, Liang
    Luan, Zhenhua
    Huang, Weiping
    SYMMETRY-BASEL, 2019, 11 (07):
  • [5] Fast binary feature selection with conditional mutual information
    Fleuret, F
    JOURNAL OF MACHINE LEARNING RESEARCH, 2004, 5 : 1531 - 1555
  • [6] Feature selection based on weighted conditional mutual information
    Zhou, Hongfang
    Wang, Xiqian
    Zhang, Yao
    APPLIED COMPUTING AND INFORMATICS, 2024, 20 (1/2) : 55 - 68
  • [7] Fast binary feature selection with conditional mutual information
    Fleuret, Francois
    Journal of Machine Learning Research, 2004, 5 : 1531 - 1555
  • [8] Gene Selection by Mutual Nearest Neighbor Approach
    Shashirekha, H. L.
    Wani, Agar Hussain
    2015 INTERNATIONAL CONFERENCE ON EMERGING RESEARCH IN ELECTRONICS, COMPUTER SCIENCE AND TECHNOLOGY (ICERECT), 2015, : 398 - 402
  • [9] FEATURE SELECTION ALGORITHM BASED ON CONDITIONAL DYNAMIC MUTUAL INFORMATION
    Wang Liping
    INTERNATIONAL JOURNAL ON SMART SENSING AND INTELLIGENT SYSTEMS, 2015, 8 (01): : 316 - 337
  • [10] Conditional mutual information based feature selection for classification task
    Novovicova, Jana
    Somol, Petr
    Haindl, Michal
    Pudil, Pavel
    PROGRESS IN PATTERN RECOGNITION, IMAGE ANALYSIS AND APPLICATIONS, PROCEEDINGS, 2007, 4756 : 417 - 426