Unsupervised feature selection via discrete spectral clustering and feature weights

被引:14
作者
Shang, Ronghua [1 ]
Kong, Jiarui [1 ]
Wang, Lujuan [1 ]
Zhang, Weitong [1 ]
Wang, Chao [2 ]
Li, Yangyang [1 ]
Jiao, Licheng [1 ]
机构
[1] Xidian Univ, Sch Artificial Intelligence, Key Lab Intelligent Percept & Image Understanding, Minist Educ, Xian 710071, Shanxi, Peoples R China
[2] Zhejiang Lab, Res Ctr Big Data Intelligence, Hangzhou 311121, Zhejiang, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Unsupervised feature selection; Discrete spectral clustering; Feature weights; Orthogonal regression; SPARSE REGRESSION; FEATURE SUBSET; FACTORIZATION; FRAMEWORK;
D O I
10.1016/j.neucom.2022.10.053
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Most of the existing unsupervised feature selection methods learn the cluster structure through spectral clustering, and then use various regression models to introduce the data matrix into the indicator matrix to obtain feature selection matrix. In these methods, the clustering indicator matrix is usually continuous value, which is not the best choice for the matrix in terms of its supervising role in feature selection. Based on this, unsupervised feature selection via discrete spectral clustering and feature weights (FSDSC) is proposed in this paper. First, FSDSC integrates regression model and spectral clustering in a unified framework for feature selection, and introduces a feature weight matrix, which intuitively expresses the importance of each feature with its diagonal elements. Compared with the common feature selection matrix that requires constraints such as sparse regular items, the appearance of the feature weight matrix reduces the complexity of the model and simplifies the calculation process of feature eval-uation. Secondly, for the value of the indicators matrix, the spectral clustering is improved to obtain a discrete clustering indicator matrix, which provides clearer guidance information for feature selection. Finally, in order to avoid trivial solutions, the transformation matrix is constrained by orthogonal con-straint. The combination of the orthogonal regression model and spectral clustering enables the algo-rithm to perform feature selection and manifold information learning at the same time, thereby preserving the local geometric structure of data. Compared with other excellent unsupervised feature selection algorithms, the experimental results prove the effectiveness of the proposed algorithm.(c) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页码:106 / 117
页数:12
相关论文
共 59 条
  • [1] Belkin M, 2002, ADV NEUR IN, V14, P585
  • [2] Cai D., 2010, P 16 ACM SIGKDD INT, P333, DOI DOI 10.1145/1835804.1835848
  • [3] Feature selection in machine learning: A new perspective
    Cai, Jie
    Luo, Jiawei
    Wang, Shulin
    Yang, Sheng
    [J]. NEUROCOMPUTING, 2018, 300 : 70 - 79
  • [4] The Fisher-Markov Selector: Fast Selecting Maximally Separable Feature Subset for Multiclass Classification with Applications to High-Dimensional Data
    Cheng, Qiang
    Zhou, Hongbo
    Cheng, Jie
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2011, 33 (06) : 1217 - 1233
  • [5] Minimum redundancy feature selection from microarray gene expression data
    Ding, C
    Peng, HC
    [J]. PROCEEDINGS OF THE 2003 IEEE BIOINFORMATICS CONFERENCE, 2003, : 523 - 528
  • [6] Dy JG, 2004, J MACH LEARN RES, V5, P845
  • [7] Locality and similarity preserving embedding for feature selection
    Fang, Xiaozhao
    Xu, Yong
    Li, Xuelong
    Fan, Zizhu
    Liu, Hong
    Chen, Yan
    [J]. NEUROCOMPUTING, 2014, 128 : 304 - 315
  • [8] Fanhua Shang, 2011, Proceedings of the 2011 IEEE 11th International Conference on Data Mining (ICDM 2011), P597, DOI 10.1109/ICDM.2011.89
  • [9] Gibbons JD., 2014, Nonparametric statistical inference, Fourth edition: Revised and expanded
  • [10] He X., 2003, Locality Preserving Projections