SMEM: A Subspace Merging Based Evolutionary Method for High-Dimensional Feature Selection

被引:0
|
作者
Li, Kaixuan [1 ]
Jiang, Shibo [2 ]
Zhang, Rui [3 ]
Qiu, Jianfeng [2 ]
Zhang, Lei [3 ]
Yang, Lixia [4 ]
Cheng, Fan [2 ]
机构
[1] Anhui Univ, Informat Mat & Intelligent Sensing Lab Anhui Prov, Hefei 230601, Peoples R China
[2] Anhui Univ, Inst Informat Mat & Intelligent Sensing Lab Anhui, Sch Artificial Intelligence, Hefei 230601, Peoples R China
[3] Anhui Univ, Sch Comp Sci & Technol, Hefei 230601, Peoples R China
[4] Anhui Univ, Inst Informat Mat & Intelligent Sensing Lab Anhui, Sch Elect & Informat Engn, Hefei 230601, Peoples R China
来源
IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE | 2024年
基金
中国国家自然科学基金;
关键词
Feature extraction; Search problems; Merging; Optimization; Sorting; History; Fans; High-dimensional feature selection; multi-objective evolutionary optimization; subspace division; pairwise subspace merging; BINARY DIFFERENTIAL EVOLUTION; PARTICLE SWARM OPTIMIZATION; ARTIFICIAL BEE COLONY; GENETIC ALGORITHM; CLASSIFICATION;
D O I
10.1109/TETCI.2024.3451695
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In the past decade, evolutionary algorithms (EAs) have shown their promising performance in solving the problem of feature selection. Despite that, it is still quite challenging to design the EAs for high-dimensional feature selection (HDFS), since the increasing number of features causes the search space of EAs grows exponentially, which is known as the "curse of dimensionality". To tackle the issue, in this paper, a Subspace Merging based Evolutionary Method, termed SMEM is suggested. In SMEM, to avoid directly optimizing the large search space of HDFS, the original feature space of HDFS is firstly divided into several independent low-dimensional subspaces. In each subspace, a subpopulation is evolved to obtain the latent good feature subsets quickly. Then, to avoid some features being missed, these low-dimensional subspaces merge in pairs, and the further search is carried on the merged subspaces. During the evolving of each merged subspace, the good feature subsets obtained from previous subspace pair are fully utilized. The above subspace merging procedure repeats, and the performance of SMEM is improved gradually, until in the end, all the subspaces are merged into one final space. At that time, the final space is also the original feature space in HDFS, which ensures all the features in the data is considered. Experimental results on different high-dimensional datasets demonstrate the effectiveness and the efficiency of the proposed SMEM, when compared with the state-of-the-arts.
引用
收藏
页数:16
相关论文
共 50 条
  • [21] High-dimensional data clustering using k-means subspace feature selection
    Wang, Xiao-Dong
    Chen, Rung-Ching
    Yan, Fei
    Journal of Network Intelligence, 2019, 4 (03): : 80 - 87
  • [22] Feature selection for high-dimensional data
    Destrero A.
    Mosci S.
    De Mol C.
    Verri A.
    Odone F.
    Computational Management Science, 2009, 6 (1) : 25 - 40
  • [23] Feature selection for high-dimensional data
    Bolón-Canedo V.
    Sánchez-Maroño N.
    Alonso-Betanzos A.
    Progress in Artificial Intelligence, 2016, 5 (2) : 65 - 75
  • [24] Enhanced NSGA-II-based feature selection method for high-dimensional classification
    Li, Min
    Ma, Huan
    Lv, Siyu
    Wang, Lei
    Deng, Shaobo
    INFORMATION SCIENCES, 2024, 663
  • [25] An Efficient Quantification Method Based on Feature Selection for High-Dimensional Uncertainties of Multistage Compressors
    Wang, Junying
    Zheng, Xinqian
    Yang, Heli
    Sun, Zhenzhong
    Song, Zhaoyun
    Fu, Yu
    JOURNAL OF ENGINEERING FOR GAS TURBINES AND POWER-TRANSACTIONS OF THE ASME, 2023, 145 (02):
  • [26] A high-dimensional classification approach based on class-dependent feature subspace
    Chen, Fuzan
    Wu, Harris
    Dou, Runliang
    Li, Minqiang
    INDUSTRIAL MANAGEMENT & DATA SYSTEMS, 2017, 117 (10) : 2325 - 2339
  • [27] Feature selection based on geometric distance for high-dimensional data
    Lee, J. -H.
    Oh, S. -Y.
    ELECTRONICS LETTERS, 2016, 52 (06) : 473 - 474
  • [28] Scalable Feature Selection in High-Dimensional Data Based on GRASP
    Moshki, Mohsen
    Kabiri, Peyman
    Mohebalhojeh, Alireza
    APPLIED ARTIFICIAL INTELLIGENCE, 2015, 29 (03) : 283 - 296
  • [29] Roulette wheel-based level learning evolutionary algorithm for feature selection of high-dimensional data
    Ma, Huan
    Li, Min
    Lv, Siyu
    Wang, Lei
    Deng, Shaobo
    APPLIED SOFT COMPUTING, 2024, 163
  • [30] A Cost-Sensitive Feature Selection Method for High-Dimensional Data
    An, Chaojie
    Zhou, Qifeng
    14TH INTERNATIONAL CONFERENCE ON COMPUTER SCIENCE AND EDUCATION (ICCSE 2019), 2019, : 1089 - 1094