A two-stage hybrid ant colony optimization for high-dimensional feature selection

被引:111
作者
Ma, Wenping [1 ,2 ]
Zhou, Xiaobo [1 ,2 ]
Zhu, Hao [1 ,2 ]
Li, Longwei [1 ,2 ]
Jiao, Licheng [1 ,2 ]
机构
[1] Xidian Univ, Int Res Ctr Intelligent Percept & Computat, Minist Educ, Key Lab Intelligent Percept & Image Understanding, Xian 710071, Shaanxi, Peoples R China
[2] Xidian Univ, Sch Artificial Intelligence, Joint Int Res Lab Intelligent Percept & Computat, Xian 710071, Shaanxi, Peoples R China
基金
中国国家自然科学基金; 中国博士后科学基金;
关键词
Feature selection; Ant colony optimization; High-dimensional data; Classification; Optimal feature subset size; FEATURE SUBSET; ALGORITHM; CLASSIFICATION;
D O I
10.1016/j.patcog.2021.107933
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Ant colony optimization (ACO) is widely used in feature selection owing to its excellent global/local search capabilities and flexible graph representation. However, the current ACO-based feature selection methods are mainly applied to low-dimensional datasets. For thousands of dimensional datasets, the search for the optimal feature subset (OFS) becomes extremely difficult due to the exponential increase of the search space. In this paper, we propose a two-stage hybrid ACO for high-dimensional feature se-lection (TSHFS-ACO). As an additional stage, it uses the interval strategy to determine the size of OFS for the following OFS search. Compared to the traditional one-stage methods that determine the size of OFS and search for OFS simultaneously, the stage of checking the performance of partial feature number endpoints in advance helps to reduce the complexity of the algorithm and alleviate the algorithm from getting into a local optimum. Moreover, the advanced ACO algorithm embeds the hybrid model, which uses the features' inherent relevance attributes and the classification performance to guide OFS search. The test results on eleven high-dimensional public datasets show that TSHFS-ACO is suitable for high-dimensional feature selection. The obtained OFS has state-of-the-art performance on most datasets. And compared with other ACO-based feature selection methods, TSHFS-ACO has a shorter running time. (c) 2021 Elsevier Ltd. All rights reserved.
引用
收藏
页数:13
相关论文
共 43 条
[11]   Fast unfolding of communities in large networks [J].
Blondel, Vincent D. ;
Guillaume, Jean-Loup ;
Lambiotte, Renaud ;
Lefebvre, Etienne .
JOURNAL OF STATISTICAL MECHANICS-THEORY AND EXPERIMENT, 2008,
[12]  
Chen T., ery and Data Mining, P785, DOI DOI 10.1145/2939672.2939785
[13]   Efficiently solving the Traveling Thief Problem using hill climbing and simulated annealing [J].
El Yafrani, Mohamed ;
Ahiod, Belaid .
INFORMATION SCIENCES, 2018, 432 :231-244
[14]   Feature subset selection in large dimensionality domains [J].
Gheyas, Iffat A. ;
Smith, Leslie S. .
PATTERN RECOGNITION, 2010, 43 (01) :5-13
[15]   A wrapper-filter feature selection technique based on ant colony optimization [J].
Ghosh, Manosij ;
Guha, Ritam ;
Sarkar, Ram ;
Abraham, Ajith .
NEURAL COMPUTING & APPLICATIONS, 2020, 32 (12) :7839-7857
[16]   Feature selection for high-dimensional classification using a competitive swarm optimizer [J].
Gu, Shenkai ;
Cheng, Ran ;
Jin, Yaochu .
SOFT COMPUTING, 2018, 22 (03) :811-822
[17]  
Guyon Isabelle, 2003, J. Mach. Learn. Res., V3, P1157
[18]   Effective Automated Feature Construction and Selection for Classification of Biological Sequences [J].
Kamath, Uday ;
De Jong, Kenneth ;
Shehu, Amarda .
PLOS ONE, 2014, 9 (07)
[19]   An improved feature selection method based on ant colony optimization (ACO) evaluated on face recognition system [J].
Kanan, Hamidreza Rashidy ;
Faez, Karim .
APPLIED MATHEMATICS AND COMPUTATION, 2008, 205 (02) :716-725
[20]   An advanced ACO algorithm for feature subset selection [J].
Kashef, Shima ;
Nezamabadi-pour, Hossein .
NEUROCOMPUTING, 2015, 147 :271-279