Feature subset selection with cumulate conditional mutual information minimization

被引:39
作者
Zhang, Yishi [1 ]
Zhang, Zigang [1 ]
机构
[1] Huazhong Univ Sci & Technol, Sch Management, Wuhan 430074, Peoples R China
基金
美国国家科学基金会;
关键词
Classification; Feature selection; Conditional mutual information; Relevance; Redundancy; CLASSIFICATION; DEPENDENCY; RELEVANCE;
D O I
10.1016/j.eswa.2011.12.003
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Feature selection is one of the core issues in designing pattern recognition and machine learning systems, and has attracted considerable attention in the literature. In this paper, a new feature subset selection algorithm with conditional mutual information is proposed, which firstly guarantees to find a subset of which the mutual information with the class is the same as that of the original set of features, and then eliminates potential redundant features from the view of minimal information loss based on the cumulate conditional mutual information minimization criterion. From the reliability point of view, this criterion can also abate the disturbance caused by sample insufficiency in conditional mutual information estimation. In addition, a fast implementation of conditional mutual information estimation is proposed and used to tackle the computationally intractable problem. Empirical results verify that our algorithm is efficient and achieves better accuracy than several representative feature selection algorithms for three typical classifiers on various datasets. (C) 2011 Elsevier Ltd. All rights reserved.
引用
收藏
页码:6078 / 6088
页数:11
相关论文
共 46 条
[1]  
AHA DW, 1991, MACH LEARN, V6, P37, DOI 10.1007/BF00153759
[2]  
Aliferis CF, 2010, J MACH LEARN RES, V11, P171
[3]   On the approximability of minimizing nonzero variables or unsatisfied relations in linear systems [J].
Amaldi, E ;
Kann, V .
THEORETICAL COMPUTER SCIENCE, 1998, 209 (1-2) :237-260
[4]  
[Anonymous], 1991, ELEMENTS INFORM THEO, DOI [DOI 10.1002/0471200611, 10.1002/0471200611]
[5]  
[Anonymous], 2000, DATA MINING PRACTICA
[6]   USING MUTUAL INFORMATION FOR SELECTING FEATURES IN SUPERVISED NEURAL-NET LEARNING [J].
BATTITI, R .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1994, 5 (04) :537-550
[7]  
Bellman R., 1961, Adaptive Control Processes: A Guided Tour, DOI DOI 10.1515/9781400874668
[8]   Selection of relevant features and examples in machine learning [J].
Blum, AL ;
Langley, P .
ARTIFICIAL INTELLIGENCE, 1997, 97 (1-2) :245-271
[9]   BEST 2 INDEPENDENT MEASUREMENTS ARE NOT 2 BEST [J].
COVER, TM .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS, 1974, SMC4 (01) :116-117
[10]  
Dash M., 1997, Intelligent Data Analysis, V1