Fast Bi-Objective Feature Selection Using Entropy Measures and Bayesian Inference

被引:3
作者
Mei, Yi [1 ]
Xue, Bing [1 ]
Zhang, Mengjie [1 ]
机构
[1] Victoria Univ Wellington, Sch Engn & CS, Wellington, New Zealand
来源
GECCO'16: PROCEEDINGS OF THE 2016 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE | 2016年
关键词
Feature Selection; Multi-Objective Computation; Generalization; MUTUAL INFORMATION; OPTIMIZATION; ALGORITHM; CLASSIFICATION; SIMILARITY;
D O I
10.1145/2908812.2908823
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
The entropy measures have been used in feature selection for decades, and showed competitive performance. In general, the problem aims at minimizing the conditional entropy of the class label on the selected features. However, the generalization of the entropy measures has been neglected in literature. Specifically, the use of conditional entropy has two critical issues. First, the empirical conditional distribution of the class label may have a low confidence and thus is unreliable. Second, there may not be enough training instances for the selected features, and it is highly likely to encounter new examples in the test set. To address these issues, a bi-objective optimization model with a modified entropy measure called the Bayesian entropy is proposed. This model considers the confidence of the optimized conditional entropy value as well as the conditional entropy value itself. As a result, it produces multiple feature subsets with different trade-offs between the entropy value and its confidence. The experimental results demonstrate that by solving the proposed optimization model with the new entropy measure, the number of features can be dramatically reduced within a much shorter time than the existing algorithms. Furthermore, similar or even better classification accuracy was achieved for most test problems.
引用
收藏
页码:469 / 476
页数:8
相关论文
共 28 条
[1]  
[Anonymous], INT J SYST SCI
[2]  
Bache K., 2013, UCI Machine Learning Repository
[3]  
Bobrowski L., 1988, 9th International Conference on Pattern Recognition (IEEE Cat. No.88CH2614-6), P544, DOI 10.1109/ICPR.1988.28287
[4]  
Cervante L, 2013, LECT NOTES COMPUT SC, V7832, P25, DOI 10.1007/978-3-642-37198-1_3
[5]  
Cervante Liam., 2012, 2012 IEEE Congress on Evolutionary Computation, P1
[6]  
Dash M., 1997, Intelligent Data Analysis, V1
[7]   A fast and elitist multiobjective genetic algorithm: NSGA-II [J].
Deb, K ;
Pratap, A ;
Agarwal, S ;
Meyarivan, T .
IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2002, 6 (02) :182-197
[8]   Feature Selection With Harmony Search [J].
Diao, Ren ;
Shen, Qiang .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS, 2012, 42 (06) :1509-1523
[9]   Normalized Mutual Information Feature Selection [J].
Estevez, Pablo. A. ;
Tesmer, Michel ;
Perez, Claudio A. ;
Zurada, Jacek A. .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 2009, 20 (02) :189-201
[10]   Research on collaborative negotiation for e-commerce. [J].
Feng, YQ ;
Lei, Y ;
Li, Y ;
Cao, RZ .
2003 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND CYBERNETICS, VOLS 1-5, PROCEEDINGS, 2003, :2085-2088