Theoretical evaluation of feature selection methods based on mutual information

被引:51
作者
Pascoal, Claudia [1 ,2 ]
Oliveira, M. Rosario [1 ,2 ]
Pacheco, Antonio [1 ,2 ]
Valadas, Rui [3 ,4 ]
机构
[1] Univ Lisbon, Inst Super Tecn, CEMAT, Ave Rovisco Pais, P-1049001 Lisbon, Portugal
[2] Univ Lisbon, Inst Super Tecn, Dept Math, Ave Rovisco Pais, P-1049001 Lisbon, Portugal
[3] Univ Lisbon, Inst Super Tecn, IT, P-1049001 Lisbon, Portugal
[4] Univ Lisbon, Inst Super Tecn, Dept Elect & Comp Engn, P-1049001 Lisbon, Portugal
关键词
Feature selection; Mutual information; Entropy; RELEVANCE;
D O I
10.1016/j.neucom.2016.11.047
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Feature selection methods are usually evaluated by wrapping specific classifiers and datasets in the evaluation process, resulting very often in unfair comparisons between methods. Iii this work, we develop a theoretical framework that allows obtaining the true feature ordering of two-dimensional sequential forward feature selection methods based on mutual information, which is independent of entropy or mutual information estimation methods, classifiers, or datasets, and leads to an undoubtful comparison of the methods. Moreover, the theoretical framework unveils problems intrinsic to some methods that are otherwise difficult to detect, namely inconsistencies in the construction of the objective function used to select the candidate features, due to various types of indeterminations and to the possibility of the entropy of continuous random variables taking null and negative values.
引用
收藏
页码:168 / 181
页数:14
相关论文
共 50 条
[31]   Theoretical and empirical study on the potential inadequacy of mutual information for feature selection in classification [J].
Frenay, Benoit ;
Doquire, Gauthier ;
Verleysen, Michel .
NEUROCOMPUTING, 2013, 112 :64-78
[32]   Feature Selection Based on Mutual Information and Its Application in Hyperspectral Image Classification [J].
Yao, Na ;
Lin, Zongjian ;
Zhang, Jingxiong .
KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, 2010, 6291 :561-566
[33]   A Feature Subset Selection Method Based On High-Dimensional Mutual Information [J].
Zheng, Yun ;
Kwoh, Chee Keong .
ENTROPY, 2011, 13 (04) :860-901
[34]   Feature selection using mutual information based uncertainty measures for tumor classification [J].
Sun, Lin ;
Xu, Jiucheng .
BIO-MEDICAL MATERIALS AND ENGINEERING, 2014, 24 (01) :763-770
[35]   A New Approach for Feature Selection from Microarray Data Based on Mutual Information [J].
Tang, Jian ;
Zhou, Shuigeng .
IEEE-ACM TRANSACTIONS ON COMPUTATIONAL BIOLOGY AND BIOINFORMATICS, 2016, 13 (06) :1004-1015
[36]   Low bias histogram-based estimation of mutual information for feature selection [J].
Hacine-Gharbi, Abdenour ;
Ravier, Philippe ;
Harba, Rachid ;
Mohamadi, Tayeb .
PATTERN RECOGNITION LETTERS, 2012, 33 (10) :1302-1308
[37]   Feature selection based on quality of information [J].
Liu, Jinghua ;
Lin, Yaojin ;
Lin, Menglei ;
Wu, Shunxiang ;
Zhang, Jia .
NEUROCOMPUTING, 2017, 225 :11-22
[38]   Feature selection using Joint Mutual Information Maximisation [J].
Bennasar, Mohamed ;
Hicks, Yulia ;
Setchi, Rossitza .
EXPERT SYSTEMS WITH APPLICATIONS, 2015, 42 (22) :8520-8532
[39]   Feature Selection for Bearing Fault Detection Based on Mutual Information [J].
Kappaganthu, Karthik ;
Nataraj, C. ;
Samanta, Biswanath .
IUTAM SYMPOSIUM ON EMERGING TRENDS IN ROTOR DYNAMICS, 2011, 25 :523-533
[40]   Fuzzy Mutual Information Feature Selection Based on Representative Samples [J].
Salem, Omar A. M. ;
Wang, Liwei .
INTERNATIONAL JOURNAL OF SOFTWARE INNOVATION, 2018, 6 (01) :58-72