A multi-objective algorithm for multi-label filter feature selection problem

被引:22
作者
Dong, Hongbin [1 ]
Sun, Jing [1 ]
Li, Tao [1 ]
Ding, Rui [2 ]
Sun, Xiaohang [1 ]
机构
[1] Harbin Engn Univ, Dept Comp Sci & Technol, Harbin 150001, Heilongjiang, Peoples R China
[2] Mudanjiang Normal Univ, Dept Comp Sci & Technol, Mudanjiang 157000, Heilongjiang, Peoples R China
基金
美国国家科学基金会;
关键词
Feature selection; Multi-objective optimization; Multi-label; PSO; PARTICLE SWARM OPTIMIZATION; FEATURE SUBSET-SELECTION; DIFFERENTIAL EVOLUTION; GENETIC ALGORITHM; MUTUAL INFORMATION; HYBRID APPROACH; CLASSIFICATION; PSO; MUTATION; SCORE;
D O I
10.1007/s10489-020-01785-2
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Feature selection is an important data preprocessing method before classification. Multi-objective optimization algorithms have been proved an effective way to solve feature selection problems. However, there are few studies on multi-objective optimization feature selection methods for multi-label data. In this paper, a multi-objective multi-label filter feature selection algorithm based on two particle swarms (MOMFS) is proposed. We use mutual information to measure the relevance between features and label sets, and the redundancy between features, which are taken as two objectives. In order to avoid Particle Swarm Optimization (PSO) from falling into the local optimum and obtaining a false Pareto front, we employ two swarms to optimize the two objectives separately and propose an improved hybrid topology based on particle's fitness value. Furthermore, an archive maintenance strategy is introduced to maintain the distribution of archive. In order to study the effectiveness of the proposed algorithm, we select five multi-label evaluation criteria and perform experiments on seven multi-label data sets. MOMFS is compared with classic single-objective multi-label feature selection algorithms, multi-objective filter and wrapper feature selection algorithms. The experimental results show that MOMFS can effectively reduce the multi-label data dimension and perform better than other approaches on five evaluation criteria.
引用
收藏
页码:3748 / 3774
页数:27
相关论文
共 64 条
[31]  
Oramas S., 2017, P INT SOC MUSIC INFO, P23
[32]   Feature selection based on mutual information: Criteria of max-dependency, max-relevance, and min-redundancy [J].
Peng, HC ;
Long, FH ;
Ding, C .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2005, 27 (08) :1226-1238
[33]  
Raquel CR, 2005, GECCO 2005: GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE, VOLS 1 AND 2, P257
[34]   Gamers' involvement detection from EEG data with cGAAM - A method for feature selection for clustering [J].
Rejer, Izabela ;
Twardochleb, Michal .
EXPERT SYSTEMS WITH APPLICATIONS, 2018, 101 :196-204
[35]   A MATHEMATICAL THEORY OF COMMUNICATION [J].
SHANNON, CE .
BELL SYSTEM TECHNICAL JOURNAL, 1948, 27 (04) :623-656
[36]   MODE: multiobjective differential evolution for feature selection and classifier ensemble [J].
Sikdar, Utpal Kumar ;
Ekbal, Asif ;
Saha, Sriparna .
SOFT COMPUTING, 2015, 19 (12) :3529-3549
[37]   Quadratic Mutual Information Feature Selection [J].
Sluga, Davor ;
Lotric, Uros .
ENTROPY, 2017, 19 (04)
[38]  
Spolaor N., 2018, APPL ARTIF INTELL, V31, P1
[39]   A Comparison of Multi-label Feature Selection Methods using the Problem Transformation Approach [J].
Spolaor, Newton ;
Cherman, Everton Alvares ;
Monard, Maria Carolina ;
Lee, Huei Diana .
ELECTRONIC NOTES IN THEORETICAL COMPUTER SCIENCE, 2013, 292 :135-151
[40]   Using PPI network autocorrelation in hierarchical multi-label classification trees for gene function prediction [J].
Stojanova, Daniela ;
Ceci, Michelangelo ;
Malerba, Donato ;
Dzeroski, Saso .
BMC BIOINFORMATICS, 2013, 14