An interactive feature selection method based on multi-step state transition algorithm for high-dimensional data

被引:7
作者
Du, Yangyi [1 ]
Zhou, Xiaojun [1 ]
Yang, Chunhua [1 ]
Huang, Tingwen [2 ]
机构
[1] Cent South Univ, Sch Automat, Changsha 410083, Peoples R China
[2] Texas A&M Univ Qatar, Doha 10587, Qatar
关键词
Feature selection; State transition algorithm; Interactive learning; Mutual information; OPTIMIZATION;
D O I
10.1016/j.knosys.2023.111102
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Feature selection (FS) has been extensively employed in classification tasks to reduce data dimensionality and enhance prediction performance effectively. Recently, hybrid filter-wrapper methods have exhibited promising results in FS problems by leveraging both advantages. However, the inadequate integration of the filter method into the wrapper method leads to the hybrid algorithms exhibiting poor efficiency in classification as datasets grow in complexity. In this paper, an interactive feature selection framework based on state transition algorithm (STA) is proposed to address high-dimensional FS problems. In this framework, prior knowledge of the features is formed via the mutual information-based filter method. The STA is a powerful search engine that traverses the feature space in the wrapper stage. Moreover, an external trainer with prior knowledge will guide the exploration direction of STA in a simple but efficient way to accelerate the search process. And a self-adaptive mechanism is proposed to adjust the prior knowledge during the search process. Specifically, the external trainer establishes the interactive loop, while the self-adaptive mechanism aims to feed feature information back to the loop. In addition, the search process is prevented from being trapped in local optima by employing a multi-step STA, which allows for continuous transformations of the solution in probability. Finally, the proposed FS method is applied to various public classification datasets. The experimental results demonstrate that the proposed method is a highly competitive FS method, outperforming several state-of-the-art algorithms in generating an optimal subset of features.
引用
收藏
页数:15
相关论文
共 56 条
[1]   Gazelle optimization algorithm: a novel nature-inspired metaheuristic optimizer [J].
Agushaka, Jeffrey O. ;
Ezugwu, Absalom E. ;
Abualigah, Laith .
NEURAL COMPUTING & APPLICATIONS, 2023, 35 (05) :4099-4131
[2]   Dwarf Mongoose Optimization Algorithm [J].
Agushaka, Jeffrey O. ;
Ezugwu, Absalom E. ;
Abualigah, Laith .
COMPUTER METHODS IN APPLIED MECHANICS AND ENGINEERING, 2022, 391
[3]  
Ahadzadeh B, 2023, IEEE Transactions on Evolutionary Computation
[4]   A fast Branch-and-Bound algorithm for U-curve feature selection [J].
Atashpaz-Gargari, Esmaeil ;
Reis, Marcelo S. ;
Braga-Neto, Ulisses M. ;
Barrera, Junior ;
Dougherty, Edward R. .
PATTERN RECOGNITION, 2018, 73 :172-188
[5]   A survey on swarm intelligence approaches to feature selection in data mining [J].
Bach Hoai Nguyen ;
Xue, Bing ;
Zhang, Mengjie .
SWARM AND EVOLUTIONARY COMPUTATION, 2020, 54
[6]  
Tran B, 2016, IEEE C EVOL COMPUTAT, P3801, DOI 10.1109/CEC.2016.7744271
[7]  
Blake CL, 1998, Uci repository of machine learning databases
[8]   A Steering-Matrix-Based Multiobjective Evolutionary Algorithm for High-Dimensional Feature Selection [J].
Cheng, Fan ;
Chu, Feixiang ;
Xu, Yi ;
Zhang, Lei .
IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (09) :9695-9708
[9]   Gene selection and classification using Taguchi chaotic binary particle swarm optimization [J].
Chuang, Li-Yeh ;
Yang, Cheng-San ;
Wu, Kuo-Chuan ;
Yang, Cheng-Hong .
EXPERT SYSTEMS WITH APPLICATIONS, 2011, 38 (10) :13367-13377
[10]   Clustering cancer gene expression data: a comparative study [J].
de Souto, Marcilio C. P. ;
Costa, Ivan G. ;
de Araujo, Daniel S. A. ;
Ludermir, Teresa B. ;
Schliep, Alexander .
BMC BIOINFORMATICS, 2008, 9 (1)