Simultaneous Feature Selection and Support Vector Machine Optimization Using the Grasshopper Optimization Algorithm

被引:171
作者
Aljarah, Ibrahim [1 ]
Al-Zoubi, Ala M. [1 ]
Faris, Hossam [1 ]
Hassonah, Mohammad A. [1 ]
Mirjalili, Seyedali [2 ]
Saadeh, Heba [3 ]
机构
[1] Univ Jordan, Business Informat Technol Dept, King Abdullah II Sch Informat Technol, Amman, Jordan
[2] Griffith Univ, Sch Informat & Commun Technol, Brisbane, Qld 4111, Australia
[3] Univ Jordan, Dept Comp Sci, King Abdullah II Sch Informat Technol, Amman, Jordan
关键词
SVM; Support vector machine; Grasshopper optimization algorithm; GOA; Optimisation; Feature selection; Metaheuristics; PARTICLE SWARM OPTIMIZATION; ANOMALY DETECTION; CLASSIFICATION; ROBUST;
D O I
10.1007/s12559-017-9542-9
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Support vector machine (SVM) is considered to be one of the most powerful learning algorithms and is used for a wide range of real-world applications. The efficiency of SVM algorithm and its performance mainly depends on the kernel type and its parameters. Furthermore, the feature subset selection that is used to train the SVM model is another important factor that has a major influence on it classification accuracy. The feature subset selection is a very important step in machine learning, specially when dealing with high-dimensional data sets. Most of the previous researches handled these important factors separately. In this paper, we propose a hybrid approach based on the Grasshopper optimisation algorithm (GOA), which is a recent algorithm inspired by the biological behavior shown in swarms of grasshoppers. The goal of the proposed approach is to optimize the parameters of the SVM model, and locate the best features subset simultaneously. Eighteen low- and high-dimensional benchmark data sets are used to evaluate the accuracy of the proposed approach. For verification, the proposed approach is compared with seven well-regarded algorithms. Furthermore, the proposed approach is compared with grid search, which is the most popular technique for tuning SVM parameters. The experimental results show that the proposed approach outperforms all of the other techniques in most of the data sets in terms of classification accuracy, while minimizing the number of selected features.
引用
收藏
页码:478 / 495
页数:18
相关论文
共 60 条
[1]  
Aljarah I., 2016, Neural Computing and Applications, P1
[2]   Optimizing connection weights in neural networks using the whale optimization algorithm [J].
Aljarah, Ibrahim ;
Faris, Hossam ;
Mirjalili, Seyedali .
SOFT COMPUTING, 2018, 22 (01) :1-15
[3]  
[Anonymous], NEURAL COMPUT APPL
[4]  
[Anonymous], 2006, P IEEE C COMPUTER VI, DOI DOI 10.1109/CVPR.2006.301
[5]  
[Anonymous], 2009, NABIC 2009 WORLD C 2
[6]  
[Anonymous], 2011, ACM T INTEL SYST TEC, DOI DOI 10.1145/1961189.1961199
[7]   Support Vector Machines Trained with Evolutionary Algorithms Employing Kernel Adatron for Large Scale Classification of Protein Structures [J].
Arana-Daniel, Nancy ;
Gallegos, Alberto A. ;
Lopez-Franco, Carlos ;
Alanis, Alma Y. ;
Morales, Jacob ;
Lopez-Franco, Adriana .
EVOLUTIONARY BIOINFORMATICS, 2016, 12 :285-302
[8]   Effects of principle component analysis on assessment of coronary artery diseases using support vector machine [J].
Babaoglu, Ismail ;
Findik, Oguz ;
Bayrak, Mehmet .
EXPERT SYSTEMS WITH APPLICATIONS, 2010, 37 (03) :2182-2185
[9]   A PSO and pattern search based memetic algorithm for SVMs parameters optimization [J].
Bao, Yukun ;
Hu, Zhongyi ;
Xiong, Tao .
NEUROCOMPUTING, 2013, 117 :98-106
[10]  
Beyer K, 1999, LECT NOTES COMPUT SC, V1540, P217