Feature selection by integrating two groups of feature evaluation criteria

被引:54
作者
Gao, Wanfu [1 ]
Hu, Liang [1 ]
Zhang, Ping [2 ]
Wang, Feng [1 ]
机构
[1] Jilin Univ, Coll Comp Sci & Technol, Changchun 130012, Jilin, Peoples R China
[2] Jilin Univ, Coll Software, Changchun 130012, Jilin, Peoples R China
基金
中国国家自然科学基金;
关键词
Feature selection; Information theory; Classification; Class-independent feature redundancy; Class-dependent feature redundancy; MUTUAL INFORMATION; RELEVANCE; DEPENDENCY;
D O I
10.1016/j.eswa.2018.05.029
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Feature selection is a preprocessing step in many application areas that are relevant to expert and intelligent systems, such as data mining and machine learning. Feature selection criteria that are based on information theory can be generally sorted into two categories. The criteria in the first group focus on minimizing feature redundancy, whereas those in the second group aim to maximize new classification information. However, both groups of feature evaluation criteria fail to balance the importance of feature redundancy and new classification information. Therefore, we propose a hybrid feature selection method named Minimal Redundancy-Maximal New Classification Information (MR-MNCI) that integrates the two groups of feature selection criteria. Moreover, according to the characteristics of the two groups of selection criteria, we adopt class-dependent feature redundancy and class-independent feature redundancy. To evaluate MR-MNCI, seven competing feature selection methods are compared with our method on 12 real-world data sets. Our method achieves the best classification performance in terms of average classification accuracy and highest classification accuracy. (C) 2018 Elsevier Ltd. All rights reserved.
引用
收藏
页码:11 / 19
页数:9
相关论文
共 34 条
[1]   Variable Global Feature Selection Scheme for automatic classification of text documents [J].
Agnihotri, Deepak ;
Verma, Kesari ;
Tripathi, Priyanka .
EXPERT SYSTEMS WITH APPLICATIONS, 2017, 81 :268-281
[2]  
[Anonymous], 2016, ARXIV160107996
[3]  
[Anonymous], 2012, ELEMENTS INFORM THEO
[4]  
[Anonymous], 2010, P 27 INT C MACHINE L
[5]   USING MUTUAL INFORMATION FOR SELECTING FEATURES IN SUPERVISED NEURAL-NET LEARNING [J].
BATTITI, R .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1994, 5 (04) :537-550
[6]   Feature selection using Joint Mutual Information Maximisation [J].
Bennasar, Mohamed ;
Hicks, Yulia ;
Setchi, Rossitza .
EXPERT SYSTEMS WITH APPLICATIONS, 2015, 42 (22) :8520-8532
[7]   Feature Interaction Maximisation [J].
Bennasar, Mohamed ;
Setchi, Rossitza ;
Hicks, Yulia .
PATTERN RECOGNITION LETTERS, 2013, 34 (14) :1630-1635
[8]   Maximum relevance minimum common redundancy feature selection for nonlinear data [J].
Che, Jinxing ;
Yang, Youlong ;
Li, Li ;
Bai, Xuying ;
Zhang, Shenghu ;
Deng, Chengzhi .
INFORMATION SCIENCES, 2017, 409 :68-86
[9]   Conditional Mutual Information-Based Feature Selection Analyzing for Synergy and Redundancy [J].
Cheng, Hongrong ;
Qin, Zhiguang ;
Feng, Chaosheng ;
Wang, Yong ;
Li, Fagen .
ETRI JOURNAL, 2011, 33 (02) :210-218
[10]   Ensemble feature selection using bi-objective genetic algorithm [J].
Das, Asit K. ;
Das, Sunanda ;
Ghosh, Arka .
KNOWLEDGE-BASED SYSTEMS, 2017, 123 :116-127