Research on Feature Selection Algorithm Based on Natural Evolution Strategy

被引:0
|
作者
Zhang X. [1 ,2 ]
Li Z.-S. [1 ,2 ]
机构
[1] College of Computer Science and Technology, Jilin University, Changchun
[2] Key Laboratory of Symbolic Computation and Knowledge Engineering, Jilin University, Ministry of Education, Changchun
来源
Ruan Jian Xue Bao/Journal of Software | 2020年 / 31卷 / 12期
基金
中国国家自然科学基金;
关键词
Competitive evolution; Cooperative co-evolution; Evolution strategy; Feature selection; High-dimensional;
D O I
10.13328/j.cnki.jos.005874
中图分类号
学科分类号
摘要
Feature selection is an NP-hard problem that aims to improve the accuracy of the model by eliminating irrelevant or redundant features to reduce model training time. Therefore, feature selection is an important data preprocessing technique in the fields of machine learning, data mining, and pattern recognition. This study proposes a new feature selection algorithm MCC-NES based on natural evolutionary strategy. Firstly, the algorithm adopts natural evolutionary strategy based on diagonal covariance matrix modeling, which adaptively adjusts parameters through gradient information. Secondly, in order to enable the algorithm to effectively deal with feature selection problems, a feature coding mechanism is introduced in the initialization phase, and combined with classification accuracy and dimensional reduction, given the new fitness function. In addition, the idea of sub-population cooperative co-evolution is introduced to solve high-dimensional data. The original problem is decomposed into relatively small sub-problems to reduce the combined effect of the original problem scale and each sub-question is solved independently, and then all sub-problems are correlated to optimize the solution to the original problem. Further, applying multiple competing evolutionary populations to enhance the exploration ability of the algorithm and designing a population restart strategy to prevent the population from falling into the local optimal solution. Finally, the proposed algorithm is compared with several traditional feature selection algorithms on some UCI public datasets. The experimental results show that the proposed algorithm can effectively complete the feature selection problem and has excellent performance compared with the classical feature selection algorithm, especially when dealing with high-dimensional data. © Copyright 2020, Institute of Software, the Chinese Academy of Sciences. All rights reserved.
引用
收藏
页码:3733 / 3752
页数:19
相关论文
共 46 条
  • [1] Guyon I, Elisseeff A., An introduction to variable and feature selection, Journal of Machine Learning Research, pp. 1157-1182, (2003)
  • [2] Liu H, Motoda H, Setiono R, Et al., Feature selection: An ever evolving frontier in data mining, Journal of Machine Learning Research, pp. 4-13, (2010)
  • [3] Dash M, Liu H., Feature selection for classification, Intelligent Data Analysis, 1, 3, pp. 131-156, (1997)
  • [4] Chandrashekar G, Sahin F., A survey on feature selection methods, Computers & Electrical Engineering, 40, 1, pp. 16-28, (2014)
  • [5] Kohavi R, John GH., Wrappers for feature subset selection, Artificial Intelligence, 97, 1-2, pp. 273-324, (1997)
  • [6] Tan KC, Teoh EJ, Yu Q, Et al., A hybrid evolutionary algorithm for attribute selection in data mining, Expert Systems with Applications, 36, 4, pp. 8616-8630, (2009)
  • [7] Bart S, Gomes CP., Hill-climbing Search, Encyclopedia of Cognitive Science, (2006)
  • [8] Pudil P, Novovicova J, Kittler J, Et al., Floating search methods in feature selection, Pattern Recognition Letters, 15, 11, pp. 1119-1125, (1994)
  • [9] Moustakidis SP, Theocharis JB., SVM-FuzCoC: A novel SVM-based feature selection method using a fuzzy complementary criterion, Pattern Recognition, 43, 11, pp. 3712-3729, (2010)
  • [10] Xue B, Zhang M, Browne WN, Et al., A survey on evolutionary computation approaches to feature selection, IEEE Trans. on Evolutionary Computation, 20, 4, pp. 606-626, (2016)