Feature selection based on rough set approach, wrapper approach, and binary whale optimization algorithm

被引:77
作者
Tawhid, Mohamed A. [1 ]
Ibrahim, Abdelmonem M. [1 ,2 ]
机构
[1] Thompson Rivers Univ, Fac Sci, Dept Math, Stat, Kamloops, BC V2C 0C8, Canada
[2] Al Azhar Univ, Fac Sci, Dept Math, Assiut Branch, Assiut, Egypt
基金
加拿大自然科学与工程研究理事会;
关键词
Feature selection; Classification; Whale optimization algorithm; Rough set theory; Wrapper approach; Logistic regression; ATTRIBUTE REDUCTION; DIFFERENTIAL EVOLUTION; SEARCH; CLASSIFICATION; CANCER; SVM;
D O I
10.1007/s13042-019-00996-5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The principle of any approach for solving feature selection problem is to find a subset of the original features. Since finding a minimal subset of the features is an NP-hard problem, it is necessary to develop and propose practical and efficient heuristic algorithms. The whale optimization algorithm is a recently developed nature-inspired meta-heuristic optimization algorithm that imitates the hunting behavior of humpback whales to solve continuous optimization problems. In this paper, we propose a novel binary whale optimization algorithm (BWOA) to solve feature selection problem. BWOA is especially desirable and appealing for feature selection problem whenever there is no heuristic information that can lead the search to the optimal minimal subset. Nonetheless, whales can find the best features as they hunt the prey. Rough set theory (RST) is one of the effective algorithms for feature selection. We use RST with BWOA as the first experiment, and in the second experiment, we use a wrapper approach with BWOA on three different classifiers for feature selection. Also, we verify the performance and the effectiveness of the proposed algorithm by performing our experiments using 32 datasets from the UCI machine learning repository and comparing the proposed algorithm with some powerful existing algorithms in the literature. Furthermore, we employ two nonparametric statistical tests, Wilcoxon Signed-Rank test, and Friedman test, at 5% significance level. Our results show that the proposed algorithm can provide an efficient tool to find a minimal subset of the features.
引用
收藏
页码:573 / 602
页数:30
相关论文
共 79 条
[1]   Feature subset selection using differential evolution and a wheel based search strategy [J].
Al-Ani, Ahmed ;
Alsukker, Akram ;
Khushaba, Rami N. .
SWARM AND EVOLUTIONARY COMPUTATION, 2013, 9 :15-26
[2]  
Alcalá-Fdez J, 2011, J MULT-VALUED LOG S, V17, P255
[3]  
[Anonymous], 2010, UCI machine learning repository
[4]  
[Anonymous], 2016, PATTERN RECOGNITION
[5]  
[Anonymous], STUDIES FUZZINESS SO
[6]  
[Anonymous], 1997, Icml
[7]   Feature selection with Intelligent Dynamic Swarm and Rough Set [J].
Bae, Changseok ;
Yeh, Wei-Chang ;
Chung, Yuk Ying ;
Liu, Sin-Long .
EXPERT SYSTEMS WITH APPLICATIONS, 2010, 37 (10) :7026-7032
[8]   Different metaheuristic strategies to solve the feature selection problem [J].
Casado Yusta, Silvia .
PATTERN RECOGNITION LETTERS, 2009, 30 (05) :525-534
[9]  
Cervante L, 2013, 2013 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION (CEC), P2428
[10]   Enhancing the classification accuracy by scatter-search-based ensemble approach [J].
Chen, Shih-Chieh ;
Lin, Shih-Wei ;
Chou, Shuo-Yan .
APPLIED SOFT COMPUTING, 2011, 11 (01) :1021-1028