Adaptive Multi-Subswarm Optimisation for Feature Selection on High-Dimensional Classification

被引:18
作者
Binh Tran [1 ,2 ]
Xue, Bing [1 ]
Zhang, Mengjie [1 ]
机构
[1] Victoria Univ Wellington, Sch Engn & Comp Sci, Wellington, New Zealand
[2] Can Tho Univ, Can Tho, Vietnam
来源
PROCEEDINGS OF THE 2019 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE (GECCO'19) | 2019年
关键词
Feature selection; Particle swarm optimisation; High-dimensional data; Classification; PARTICLE SWARM OPTIMIZATION;
D O I
10.1145/3321707.3321713
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Feature space is an important factor influencing the performance of any machine learning algorithm including classification methods. Feature selection aims to remove irrelevant and redundant features that may negatively affect the learning process especially on high-dimensional data, which usually suffers from the curse of dimensionality. Feature ranking is one of the most scalable feature selection approaches to high-dimensional problems, but most of them fail to automatically determine the number of selected features as well as detect redundancy between features. Particle swarm optimisation (PSO) is a population-based algorithm which has shown to be effective in addressing these limitations. However, its performance on high-dimensional data is still limited due to the large search space and high computation cost. This study proposes the first adaptive multi-swarm optimisation (AMSO) method for feature selection that can automatically select a feature subset of high-dimensional data more effectively and efficiently than the compared methods. The subswarms are automatically and dynamically changed based on their performance during the evolutionary process. Experiments on ten high-dimensional datasets of varying difficulties have shown that AMSO is more effective and more efficient than the compared PSO-based and traditional feature selection methods in most cases.
引用
收藏
页码:481 / 489
页数:9
相关论文
共 22 条
[11]   Large-scale attribute selection using wrappers [J].
Guetlein, Martin ;
Frank, Eibe ;
Hall, Mark ;
Karwath, Andreas .
2009 IEEE SYMPOSIUM ON COMPUTATIONAL INTELLIGENCE AND DATA MINING, 2009, :332-339
[12]  
Hall MA, 2000, CORRELATION BASED FE
[13]   Correlation feature selection based improved-Binary Particle Swarm Optimization for gene selection and cancer classification [J].
Jain, Indu ;
Jain, Vinod Kumar ;
Jain, Renu .
APPLIED SOFT COMPUTING, 2018, 62 :203-215
[14]  
Kennedy J, 1995, 1995 IEEE INTERNATIONAL CONFERENCE ON NEURAL NETWORKS PROCEEDINGS, VOLS 1-6, P1942, DOI 10.1109/icnn.1995.488968
[15]   Feature Selection: A Data Perspective [J].
Li, Jundong ;
Cheng, Kewei ;
Wang, Suhang ;
Morstatter, Fred ;
Trevino, Robert P. ;
Tang, Jiliang ;
Liu, Huan .
ACM COMPUTING SURVEYS, 2018, 50 (06)
[16]   A Micro-GA Embedded PSO Feature Selection Approach to Intelligent Facial Emotion Recognition [J].
Mistry, Kamlesh ;
Zhang, Li ;
Neoh, Siew Chin ;
Lim, Chee Peng ;
Fielding, Ben .
IEEE TRANSACTIONS ON CYBERNETICS, 2017, 47 (06) :1496-1509
[17]   A Modified Binary Particle Swarm Optimization for Selecting the Small Subset of Informative Genes From Gene Expression Data [J].
Mohamad, Mohd Saberi ;
Omatu, Sigeru ;
Deris, Safaai ;
Yoshioka, Michifumi .
IEEE TRANSACTIONS ON INFORMATION TECHNOLOGY IN BIOMEDICINE, 2011, 15 (06) :813-822
[18]   A hybrid particle swarm optimization for feature subset selection by integrating a novel local search strategy [J].
Moradi, Parham ;
Gholampour, Mozhgan .
APPLIED SOFT COMPUTING, 2016, 43 :117-130
[19]   A comprehensive evaluation of multicategory classification methods for microarray gene expression cancer diagnosis [J].
Statnikov, A ;
Aliferis, CF ;
Tsamardinos, I ;
Hardin, D ;
Levy, S .
BIOINFORMATICS, 2005, 21 (05) :631-643
[20]  
van der Maaten L, 2008, J MACH LEARN RES, V9, P2579