Feature selection using a set based discrete particle swarm optimization and a novel feature subset evaluation criterion

被引:9
作者
Qiu, Chenye [1 ]
Xiang, Fei [2 ]
机构
[1] Nanjing Univ Posts & Telecommun, Sch Internet Things, 66 Xinmofan Rd, Nanjing 210003, Jiangsu, Peoples R China
[2] Coordinat Ctr China, Natl Comp Network Emergency Response Tech Team, Beijing 100029, Peoples R China
关键词
Feature selection; particle swarm optimization; discrete search space; mutual information; feature subset evaluation criterion; MUTUAL INFORMATION; CLASSIFICATION; FRAMEWORK; RELEVANCE; VERSION;
D O I
10.3233/IDA-173735
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In many machine learning or patter recognition tasks such as classification, datasets with a large number of features are involved. Feature selection aims at eliminating the redundant and irrelevant features which would bring computational burden and degrade the performance of learning algorithms. Particle swarm optimization (PSO) has been widely used in feature selection due to its global search ability and computational efficiency. However, PSO was originally designed for continuous optimization problems and the discretization of PSO in feature selection is still a problem which needs further investigation. This paper develops a novel feature selection algorithm based on a set based discrete PSO (SPSO). SPSO employs a set based encoding scheme which makes it able to characterize the discrete search space in feature selection problem. It also redefines the velocity term and the corresponding arithmetic operators which enables it to search for the optimal feature subset in the discrete space. In addition, a novel feature subset evaluation criterion based on contribution rate is proposed as the fitness function in SPSO. The proposed criterion does not need any pre-determined parameter to keep the balance between relevance and redundancy of the feature subset. The proposed method is compared with six filter approaches and four wrapper approaches on ten well known UCI dataset and the experimental results demonstrate the proposed method is promising.
引用
收藏
页码:5 / 21
页数:17
相关论文
共 36 条
[1]  
[Anonymous], 2005, MACHINE LEARNING BAS
[2]  
[Anonymous], 1993, Proceedings of the 13th International Joint Conference on Artificial Intelligence, DOI DOI 10.1109/TKDE.2011.181
[3]  
[Anonymous], 2012, PROC SPRINGER C ARTI, DOI DOI 10.1007/978-3-642-32695-0_59
[4]   Consistency measures for feature selection [J].
Arauzo-Azofra, Antonio ;
Manuel Benitez, Jose ;
Luis Castro, Juan .
JOURNAL OF INTELLIGENT INFORMATION SYSTEMS, 2008, 30 (03) :273-292
[5]   Feature selection with Intelligent Dynamic Swarm and Rough Set [J].
Bae, Changseok ;
Yeh, Wei-Chang ;
Chung, Yuk Ying ;
Liu, Sin-Long .
EXPERT SYSTEMS WITH APPLICATIONS, 2010, 37 (10) :7026-7032
[6]   USING MUTUAL INFORMATION FOR SELECTING FEATURES IN SUPERVISED NEURAL-NET LEARNING [J].
BATTITI, R .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1994, 5 (04) :537-550
[7]   Unsupervised feature selection using an improved version of Differential Evolution [J].
Bhadra, Tapas ;
Bandyopadhyay, Sanghamitra .
EXPERT SYSTEMS WITH APPLICATIONS, 2015, 42 (08) :4042-4053
[8]  
Brown G, 2012, J MACH LEARN RES, V13, P27
[9]  
Cervante L, 2012, IEEE C EVOL COMPUTAT
[10]  
Chakraborty B, 2002, ISIE 2002: PROCEEDINGS OF THE 2002 IEEE INTERNATIONAL SYMPOSIUM ON INDUSTRIAL ELECTRONICS, VOLS 1-4, P315, DOI 10.1109/ISIE.2002.1026085