Bi-objective feature selection in high-dimensional datasets using improved binary chimp optimization algorithm

被引:4
作者
Al-qudah, Nour Elhuda A. [1 ]
Abed-alguni, Bilal H. [1 ]
Barhoush, Malek [2 ]
机构
[1] Yarmouk Univ, Dept Comp Sci, Irbid, Jordan
[2] Yarmouk Univ, Dept Informat Technol Cybersecur Program, Irbid, Jordan
基金
英国科研创新办公室;
关键词
Chimp optimization algorithm; Opposition-based learning; High-dimensional datasets; Feature selection; Levy flight; beta-Hill climbing algorithm; CUCKOO SEARCH; MODELS;
D O I
10.1007/s13042-024-02308-y
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The machine learning process in high-dimensional datasets is far more complicated than in low-dimensional datasets. In high-dimensional datasets, Feature Selection (FS) is necessary to decrease the complexity of learning. However, FS in high-dimensional datasets is a complex process that requires the combination of several search techniques. The Chimp Optimization Algorithm, known as ChOA, is a new meta-heuristic method inspired by the chimps' individual intellect and sexual incentive in cooperative hunting. It is basically employed in solving complex continuous optimization problems, while its binary version is frequently utilized in solving difficult binary optimization problems. Both versions of ChOA are subject to premature convergence and are incapable of effectively solving high-dimensional optimization problems. This paper proposes the Binary Improved ChOA Algorithm (BICHOA) for solving the bi-objective, high-dimensional FS problems (i.e., high-dimensional FS problems that aim to maximize the classifier's accuracy and minimize the number of selected features from a dataset). BICHOA improves the performance of ChOA using four new exploration and exploitation techniques. First, it employs the opposition-based learning approach to initially create a population of diverse binary feasible solutions. Second, it incorporates the L & eacute;vy mutation function in the main probabilistic update function of ChOA to boost its searching and exploring capabilities. Third, it uses an iterative exploration technique based on an exploratory local search method called the beta-hill climbing algorithm. Finally, it employs a new binary time-varying transfer function to calculate binary feasible solutions from the continuous feasible solutions generated by the update equations of the ChOA and beta-hill climbing algorithms. BICHOA's performance was assessed and compared against six machine learning classifiers, five integer programming methods, and nine efficient popular optimization algorithms using 25 real-world high-dimensional datasets from various domains. According to the overall experimental findings, BICHOA scored the highest accuracy, best objective value, and fewest selected features for each of the 25 real-world high-dimensional datasets. Besides, the reliability of the experimental findings was established using Friedman and Wilcoxon statistical tests.
引用
收藏
页码:6107 / 6148
页数:42
相关论文
共 100 条
[1]   Hybrid feature selection based on SLI and genetic algorithm for microarray datasets [J].
Abasabadi, Sedighe ;
Nematzadeh, Hossein ;
Motameni, Homayun ;
Akbari, Ebrahim .
JOURNAL OF SUPERCOMPUTING, 2022, 78 (18) :19725-19753
[2]   Parameter estimation of solar cells diode models by an improved opposition-based whale optimization algorithm [J].
Abd Elaziz, Mohamed ;
Oliva, Diego .
ENERGY CONVERSION AND MANAGEMENT, 2018, 171 :1843-1859
[3]   Spider wasp optimizer: a novel meta-heuristic optimization algorithm [J].
Abdel-Basset, Mohamed ;
Mohamed, Reda ;
Jameel, Mohammed ;
Abouhawwash, Mohamed .
ARTIFICIAL INTELLIGENCE REVIEW, 2023, 56 (10) :11675-11738
[4]   A new fusion of grey wolf optimizer algorithm with a two-phase mutation for feature selection [J].
Abdel-Basset, Mohamed ;
El-Shahat, Doaa ;
El-henawy, Ibrahim ;
de Albuquerque, Victor Hugo C. ;
Mirjalili, Seyedali .
EXPERT SYSTEMS WITH APPLICATIONS, 2020, 139
[5]  
Abed-alguni BH., 2018, JORDAN J COMPUT INFO, V4, P130, DOI DOI 10.5455/JJCIT.71-1532897697
[6]   Exploratory cuckoo search for solving single-objective optimization problems [J].
Abed-alguni, Bilal H. ;
Alawad, Noor Aldeen ;
Barhoush, Malek ;
Hammad, Rafat .
SOFT COMPUTING, 2021, 25 (15) :10167-10180
[7]   IBJA: An improved binary DJaya algorithm for feature selection [J].
Abed-alguni, Bilal H. ;
AL-Jarah, Saqer Hamzeh .
JOURNAL OF COMPUTATIONAL SCIENCE, 2024, 75
[8]   Improved Salp swarm algorithm for solving single-objective continuous optimization problems [J].
Abed-Alguni, Bilal H. ;
Paul, David ;
Hammad, Rafat .
APPLIED INTELLIGENCE, 2022, 52 (15) :17217-17236
[9]   Opposition-based sine cosine optimizer utilizing refraction learning and variable neighborhood search for feature selection [J].
Abed-alguni, Bilal H. ;
Alawad, Noor Aldeen ;
Al-Betar, Mohammed Azmi ;
Paul, David .
APPLIED INTELLIGENCE, 2023, 53 (11) :13224-13260
[10]   Advances in Sine Cosine Algorithm: A comprehensive survey [J].
Abualigah, Laith ;
Diabat, Ali .
ARTIFICIAL INTELLIGENCE REVIEW, 2021, 54 (04) :2567-2608