A novel bacterial foraging optimization algorithm for feature selection

被引:99
作者
Chen, Yu-Peng [1 ,2 ]
Li, Ying [1 ,2 ]
Wang, Gang [1 ,2 ]
Zheng, Yue-Feng [1 ,2 ]
Xu, Qian [1 ,2 ]
Fan, Jia-Hao [1 ,2 ]
Cui, Xue-Ting [1 ,2 ]
机构
[1] Jilin Univ, Coll Comp Sci & Technol, Changchun 130012, Peoples R China
[2] Jilin Univ, Minist Educ, Key Lab Symbol Computat & Knowledge Engn, Changchun, Peoples R China
基金
中国国家自然科学基金;
关键词
Bacterial foraging optimization algorithm; Feature selection; Classification; Support vector machine; SUPPORT VECTOR MACHINES; SET;
D O I
10.1016/j.eswa.2017.04.019
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Bacterial foraging optimization (BFO) algorithm is a new swarming intelligent method, which has a satisfactory performance in solving the continuous optimization problem based on the chemotaxis, swarming, reproduction and elimination-dispersal steps. However, BFO algorithm is rarely used to deal with feature selection problem. In this paper, we propose two novel BFO algorithms, which are named as adaptive chemotaxis bacterial foraging optimization algorithm (ACBFO) and improved swarming and elimination dispersal bacterial foraging optimization algorithm (ISEDBFO) respectively. Two improvements are presented in ACBFO. On the one hand, in order to solve the discrete problem, data structure of each bacterium is redefined to establish the mapping relationship between the bacterium and the feature subset. On the other hand, an adaptive method for evaluating the importance of features is designed. Therefore the primary features in feature subset are preserved. ISEDBFO is proposed based on ACBFO. ISEDBFO algorithm also includes two modifications. First, with the aim of describing the nature of cell to cell attraction-repulsion relationship more accurately, swarming representation is improved by means of introducing the hyperbolic tangent function. Second, in order to retain the primary features of eliminated bacteria, roulette technique is applied to the elimination-dispersal phase. In this study, ACBFO and ISEDBFO are tested with 10 public data sets of UCI. The performance of the proposed methods is compared with particle swarm optimization based, genetic algorithm based, simulated annealing based, ant lion optimization based, binary bat algorithm based and cuckoo search based approaches. The experimental results demonstrate that the average classification accuracy of the proposed algorithms is nearly 3 percentage points higher than other tested methods. Furthermore, the improved algorithms reduce the length of the feature subset by almost 3 in comparison to other methods. In addition, the modified methods achieve excellent performance on wilcoxon signed-rank test and sensitivity-specificity test. In conclusion, the novel BFO algorithms can provide important support for the expert and intelligent systems. (C) 2017 Elsevier Ltd. All rights reserved.
引用
收藏
页码:1 / 17
页数:17
相关论文
共 39 条
[1]   Support vector machines combined with feature selection for breast cancer diagnosis [J].
Akay, Mehmet Fatih .
EXPERT SYSTEMS WITH APPLICATIONS, 2009, 36 (02) :3240-3247
[2]   Feature selection using support vector machines and bootstrap methods for ventricular fibrillation detection [J].
Alonso-Atienza, Felipe ;
Luis Rojo-Alvarez, Jose ;
Rosado-Munoz, Alfredo ;
Vinagre, Juan J. ;
Garcia-Alberola, Arcadi ;
Camps-Valls, Gustavo .
EXPERT SYSTEMS WITH APPLICATIONS, 2012, 39 (02) :1956-1967
[3]  
[Anonymous], 2003, Self-Organization in Biological Systems
[4]  
[Anonymous], 1999, Swarm Intelligence
[5]  
[Anonymous], NEURAL INFORM PROCES
[6]  
[Anonymous], J INTELLIGENT INFORM
[7]  
[Anonymous], 1995, 1995 IEEE INT C
[8]   Two hybrid wrapper-filter feature selection algorithms applied to high-dimensional microarray experiments [J].
Apolloni, Javier ;
Leguizamon, Guillermo ;
Alba, Enrique .
APPLIED SOFT COMPUTING, 2016, 38 :922-932
[9]   METHODS OF HANDLING TIES IN WILCOXON SIGNED-RANK TEST [J].
CONOVER, WJ .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 1973, 68 (344) :985-988
[10]   Approximate statistical tests for comparing supervised classification learning algorithms [J].
Dietterich, TG .
NEURAL COMPUTATION, 1998, 10 (07) :1895-1923