A multi-objective algorithm for multi-label filter feature selection problem

被引:22
作者
Dong, Hongbin [1 ]
Sun, Jing [1 ]
Li, Tao [1 ]
Ding, Rui [2 ]
Sun, Xiaohang [1 ]
机构
[1] Harbin Engn Univ, Dept Comp Sci & Technol, Harbin 150001, Heilongjiang, Peoples R China
[2] Mudanjiang Normal Univ, Dept Comp Sci & Technol, Mudanjiang 157000, Heilongjiang, Peoples R China
基金
美国国家科学基金会;
关键词
Feature selection; Multi-objective optimization; Multi-label; PSO; PARTICLE SWARM OPTIMIZATION; FEATURE SUBSET-SELECTION; DIFFERENTIAL EVOLUTION; GENETIC ALGORITHM; MUTUAL INFORMATION; HYBRID APPROACH; CLASSIFICATION; PSO; MUTATION; SCORE;
D O I
10.1007/s10489-020-01785-2
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Feature selection is an important data preprocessing method before classification. Multi-objective optimization algorithms have been proved an effective way to solve feature selection problems. However, there are few studies on multi-objective optimization feature selection methods for multi-label data. In this paper, a multi-objective multi-label filter feature selection algorithm based on two particle swarms (MOMFS) is proposed. We use mutual information to measure the relevance between features and label sets, and the redundancy between features, which are taken as two objectives. In order to avoid Particle Swarm Optimization (PSO) from falling into the local optimum and obtaining a false Pareto front, we employ two swarms to optimize the two objectives separately and propose an improved hybrid topology based on particle's fitness value. Furthermore, an archive maintenance strategy is introduced to maintain the distribution of archive. In order to study the effectiveness of the proposed algorithm, we select five multi-label evaluation criteria and perform experiments on seven multi-label data sets. MOMFS is compared with classic single-objective multi-label feature selection algorithms, multi-objective filter and wrapper feature selection algorithms. The experimental results show that MOMFS can effectively reduce the multi-label data dimension and perform better than other approaches on five evaluation criteria.
引用
收藏
页码:3748 / 3774
页数:27
相关论文
共 64 条
  • [1] [Anonymous], 2016, CRITICAL CARE
  • [2] Asilian Bidgoli Azam, 2019, Evolutionary Multi-Criterion Optimization. 10th International Conference, EMO 2019. Proceedings: Lecture Notes in Computer Science (LNCS 11411), P553, DOI 10.1007/978-3-030-12598-1_44
  • [3] USING MUTUAL INFORMATION FOR SELECTING FEATURES IN SUPERVISED NEURAL-NET LEARNING
    BATTITI, R
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS, 1994, 5 (04): : 537 - 550
  • [4] Overview of particle swarm optimisation for feature selection in classification
    [J]. Tran, Binh (tran.binh@ecs.vuw.ac.nz), 1600, Springer Verlag (8886): : 605 - 617
  • [5] Hybrid of binary gravitational search algorithm and mutual information for feature selection in intrusion detection systems
    Bostani, Hamid
    Sheikhan, Mansour
    [J]. SOFT COMPUTING, 2017, 21 (09) : 2307 - 2324
  • [6] Coello CAC, 2002, IEEE C EVOL COMPUTAT, P1051, DOI 10.1109/CEC.2002.1004388
  • [7] Feature selection for clustering - A filter solution
    Dash, M
    Choi, K
    Scheuermann, P
    Liu, H
    [J]. 2002 IEEE INTERNATIONAL CONFERENCE ON DATA MINING, PROCEEDINGS, 2002, : 115 - 122
  • [8] A fast and elitist multiobjective genetic algorithm: NSGA-II
    Deb, K
    Pratap, A
    Agarwal, S
    Meyarivan, T
    [J]. IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2002, 6 (02) : 182 - 197
  • [9] An Improved Niching Binary Particle Swarm Optimization For Feature Selection
    Dong, Hongbin
    Sun, Jing
    Li, Tao
    Li, Lijie
    [J]. 2018 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2018, : 3571 - 3577
  • [10] A novel hybrid genetic algorithm with granular information for feature selection and optimization
    Dong, Hongbin
    Li, Tao
    Ding, Rui
    Sun, Jing
    [J]. APPLIED SOFT COMPUTING, 2018, 65 : 33 - 46