A self-adaptive multi-objective feature selection approach for classification problems

被引:23
作者
Xue, Yu [1 ,2 ]
Zhu, Haokai [1 ]
Neri, Ferrante [3 ]
机构
[1] Nanjing Univ Informat Sci & Technol, Sch Comp & Software, Nanjing, Jiangsu, Peoples R China
[2] Nanjing Univ Informat Sci & Technol, Engn Res Ctr Digital Forens, Minist Educ, Nanjing, Jiangsu, Peoples R China
[3] Univ Nottingham, Sch Comp Sci, COL Lab, Nottingham, England
基金
中国国家自然科学基金;
关键词
Feature selection; self-adaptive; multi-objective genetic algorithm; stagnation detection; classification; FEATURE SUBSET-SELECTION; GENETIC ALGORITHM; DIFFERENTIAL EVOLUTION; OPTIMIZATION ALGORITHM; NETWORK; SYSTEMS; INFORMATION; MODEL;
D O I
10.3233/ICA-210664
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In classification tasks, feature selection (FS) can reduce the data dimensionality and may also improve classification accuracy, both of which are commonly treated as the two objectives in FS problems. Many meta-heuristic algorithms have been applied to solve the FS problems and they perform satisfactorily when the problem is relatively simple. However, once the dimensionality of the datasets grows, their performance drops dramatically. This paper proposes a self-adaptive multi-objective genetic algorithm (SaMOGA) for FS, which is designed to maintain a high performance even when the dimensionality of the datasets grows. The main concept of SaMOGA lies in the dynamic selection of five different crossover operators in different evolution process by applying a self-adaptive mechanism. Meanwhile, a search stagnation detection mechanism is also proposed to prevent premature convergence. In the experiments, we compare SaMOGA with five multi-objective FS algorithms on sixteen datasets. According to the experimental results, SaMOGA yields a set of well converged and well distributed solutions on most data sets, indicating that SaMOGA can guarantee classification performance while removing many features, and the advantage over its counterparts is more obvious when the dimensionality of datasets grows.
引用
收藏
页码:3 / 21
页数:19
相关论文
共 90 条
[1]   Enhanced probabilistic neural network with local decision circles: A robust classifier [J].
Ahmadlou, Mehran ;
Adeli, Hojjat .
INTEGRATED COMPUTER-AIDED ENGINEERING, 2010, 17 (03) :197-210
[2]   A dynamic ensemble learning algorithm for neural networks [J].
Alam, Kazi Md Rokibul ;
Siddique, Nazmul ;
Adeli, Hojjat .
NEURAL COMPUTING & APPLICATIONS, 2020, 32 (12) :8675-8690
[3]  
[Anonymous], 1994, ICML
[4]   Middle-Level Features for the Explanation of Classification Systems by Sparse Dictionary Methods [J].
Apicella, A. ;
Isgro, F. ;
Prevete, R. ;
Tamburrini, G. .
INTERNATIONAL JOURNAL OF NEURAL SYSTEMS, 2020, 30 (08)
[5]  
Bache K., 2013, UCI machine learning repository
[6]   Methodology for analyzing the trade-offs associated with multi-objective optimization in transportation asset management under uncertainty [J].
Bai, Qiang ;
Miralinaghi, Mohammad ;
Labi, Samuel ;
Sinha, Kumares C. .
COMPUTER-AIDED CIVIL AND INFRASTRUCTURE ENGINEERING, 2021, 36 (04) :381-401
[7]  
Blum C, 2012, VARIANTS OF EVOLUTIONARY ALGORITHMS FOR REAL-WORLD APPLICATIONS, P1
[8]   A multi-objective genetic algorithm for simultaneous model and feature selection for support vector machines [J].
Bouraoui, Amal ;
Jamoussi, Salma ;
BenAyed, Yassine .
ARTIFICIAL INTELLIGENCE REVIEW, 2018, 50 (02) :261-281
[9]   Upper Limb Movement Classification Via Electromyographic Signals and an Enhanced Probabilistic Network [J].
Burns, Alexis ;
Adeli, Hojjat ;
Buford, John A. .
JOURNAL OF MEDICAL SYSTEMS, 2020, 44 (10)
[10]   HyperSPAM: A study on hyper-heuristic coordination strategies in the continuous domain [J].
Caraffini, Fabio ;
Neri, Ferrante ;
Epitropakis, Michael .
INFORMATION SCIENCES, 2019, 477 :186-202