New feature selection paradigm based on hyper-heuristic technique

被引:16
作者
Ibrahim, Rehab Ali [1 ,2 ]
Abd Elaziz, Mohamed [2 ]
Ewees, Ahmed A. [3 ,4 ]
El-Abd, Mohammed [5 ]
Lu, Songfeng [1 ]
机构
[1] Huazhong Univ Sci & Technol, Sch Cyber Sci & Engn, Wuhan 430074, Peoples R China
[2] Zagazig Univ, Dept Math, Fac Sci, Zagazig, Egypt
[3] Univ Bisha, Dept E Syst, Bisha 61922, Saudi Arabia
[4] Damietta Univ, Dept Comp, Damietta Governorate, Egypt
[5] Amer Univ Kuwait, Coll Engn & Appl Sci, POB 3323, Safat 13034, Kuwait
基金
中国博士后科学基金;
关键词
Meta-heuristic; Chaotic maps; Differential evolution; Opposition-based learning; Feature selection; Hyper-heuristic; GREY WOLF OPTIMIZATION; SALP SWARM ALGORITHM; NEAREST-NEIGHBOR; CLASSIFICATION; REGRESSION; EVOLUTIONARY; PREDICTION; SCHEME;
D O I
10.1016/j.apm.2021.04.018
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
Feature selection (FS) is a crucial step for effective data mining since it has largest effect on improving the performance of classifiers. This is achieved by removing the irrelevant features and using only the relevant features. Many metaheuristic approaches exist in the literature in attempt to address this problem. The performance of these approaches differ based on the settings of a number of factors including the use of chaotic maps, opposition -based learning (OBL) and the percentage of the population that OBL will be applied to, the metaheuristic (MH) algorithm adopted, the classifier utilized, and the threshold value used to convert real solutions to binary ones. However, it is not an easy task to identify the best settings for these different components in order to determine the relevant fea-tures for a specific dataset. Moreover, running extensive experiments to fine tune these settings for each and every dataset will consume considerable time. In order to mitigate this important issue, a hyper-heuristic based FS paradigm is proposed. In the proposed model, a two-stage approach is adopted to identify the best combination of these com-ponents. In the first stage, referred to as the training stage, the Differential Evolution (DE) algorithm is used as a controller for selecting the best combination of components to be used by the second stage. In the second stage, referred to as the testing stage, the received combination will be evaluated using a testing set. Empirical evaluation of the proposed framework is based on numerous experiments performed on the most popular 18 datasets from the UCI machine learning repository. Experimental results illustrates that the gener-ated generic configuration provides a better performance than eight other metaheuristic algorithms over all performance measures when applied to the UCI dataset. Moreover, The overall paradigm ranks at number one when compared against state-of-the-art algorithms. Finally, the generic configuration provides a very competitive performance for high dimen-sional datasets. (c) 2021 Elsevier Inc. All rights reserved.
引用
收藏
页码:14 / 37
页数:24
相关论文
共 111 条
  • [1] Opposition-based moth-flame optimization improved by differential evolution for feature selection
    Abd Elaziz, Mohamed
    Ewees, Ahmed A.
    Ibrahim, Rehab Ali
    Lu, Songfeng
    [J]. MATHEMATICS AND COMPUTERS IN SIMULATION, 2020, 168 (168) : 48 - 75
  • [2] A Hybrid Method of Sine Cosine Algorithm and Differential Evolution for Feature Selection
    Abd Elaziz, Mohamed E.
    Ewees, Ahmed A.
    Oliva, Diego
    Duan, Pengfei
    Xiong, Shengwu
    [J]. NEURAL INFORMATION PROCESSING, ICONIP 2017, PT V, 2017, 10638 : 145 - 155
  • [3] A new fusion of grey wolf optimizer algorithm with a two-phase mutation for feature selection
    Abdel-Basset, Mohamed
    El-Shahat, Doaa
    El-henawy, Ibrahim
    de Albuquerque, Victor Hugo C.
    Mirjalili, Seyedali
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2020, 139
  • [4] Binary Optimization Using Hybrid Grey Wolf Optimization for Feature Selection
    Al-Tashi, Qasem
    Kadir, Said Jadid Abdul
    Rais, Helmi Md
    Mirjalili, Seyedali
    Alhussian, Hitham
    [J]. IEEE ACCESS, 2019, 7 : 39496 - 39508
  • [5] AN INTRODUCTION TO KERNEL AND NEAREST-NEIGHBOR NONPARAMETRIC REGRESSION
    ALTMAN, NS
    [J]. AMERICAN STATISTICIAN, 1992, 46 (03) : 175 - 185
  • [6] Angel D., 2009, IEEE, V978, P4244
  • [7] A Novel Chaotic Interior Search Algorithm for Global Optimization and Feature Selection
    Arora, Sankalap
    Sharma, Manik
    Anand, Priyanka
    [J]. APPLIED ARTIFICIAL INTELLIGENCE, 2020, 34 (04) : 292 - 328
  • [8] A New Hybrid Algorithm Based on Grey Wolf Optimization and Crow Search Algorithm for Unconstrained Function Optimization and Feature Selection
    Arora, Sankalap
    Singh, Harpreet
    Sharma, Manik
    Sharma, Sanjeev
    Anand, Priyanka
    [J]. IEEE ACCESS, 2019, 7 : 26343 - 26361
  • [9] Binary butterfly optimization approaches for feature selection
    Arora, Sankalap
    Anand, Priyanka
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2019, 116 : 147 - 160
  • [10] Awad Asmaa Ahmed, 2020, Proceedings of the International Conference on Artificial Intelligence and Computer Vision (AICV2020). Advances in Intelligent Systems and Computing (AISC 1153), P159, DOI 10.1007/978-3-030-44289-7_16