Subspace learning for feature selection via rank revealing QR factorization: Fast feature selection

被引:5
作者
Moslemi, Amir [1 ]
Ahmadian, Arash [2 ]
机构
[1] Seneca Polytech, Sch Software Design & Data Sci, Toronto, ON, Canada
[2] Univ Toronto, Edward S Rogers Sr Dept Elect & Comp Engn, Toronto, ON M5S 1A1, Canada
关键词
Feature selection; Rank revealing QR factorization; Non-negative matrix factorization; Genetic algorithm and hybrid feature selection; UNSUPERVISED FEATURE-SELECTION; SUPERVISED FEATURE-SELECTION; MATRIX FACTORIZATION; MUTUAL INFORMATION; CLASSIFICATION; OPTIMIZATION; ALGORITHMS; APPROXIMATION; REDUCTION; PATTERNS;
D O I
10.1016/j.eswa.2024.124919
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The identification of informative and distinguishing features from high-dimensional data has gained significant attention in the field of machine learning. Recently, there has been growing interest in employing matrix factorization-based techniques, such as non-negative matrix factorization, for feature selection. The primary objective of feature selection using matrix factorization is to extract a lower-dimensional subspace that captures the essence of the original space. This study introduces a novel unsupervised feature selection technique that leverages rank revealing QR (RRQR) factorization. Compared to singular value decomposition (SVD) and nonnegative matrix factorization (NMF), RRQR is more computationally efficient. The uniqueness of this technique lies in the utilization of the permutation matrix of QR for feature selection. Additionally, we integrate QR factorization into the objective function of NMF to create a new unsupervised feature selection method. Furthermore, we propose a hybrid feature selection algorithm by combining RRQR and a Genetic algorithm. The algorithm eliminates redundant features using RRQR factorization and selects the most distinguishing subset of features using the Genetic algorithm. Experimental comparisons with state-of-the-art feature selection algorithms in supervised, unsupervised, and semi-supervised settings demonstrate the reliability and robustness of the proposed algorithm. The evaluation is conducted on eight microarray datasets using KNN, SVM, and C4.5 classifiers. The experimental results indicate that the proposed method achieves comparable performance to the state-of-the-art feature selection methods. Our empirical findings demonstrate that the proposed method exhibits a significantly lower computational cost compared to other techniques.
引用
收藏
页数:18
相关论文
共 50 条
[21]   FSMRank: Feature Selection Algorithm for Learning to Rank [J].
Lai, Han-Jiang ;
Pan, Yan ;
Tang, Yong ;
Yu, Rong .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2013, 24 (06) :940-952
[22]   UNSUPERVISED FEATURE SELECTION BY NONNEGATIVE SPARSITY ADAPTIVE SUBSPACE LEARNING [J].
Zhou, Nan ;
Cheng, Hong ;
Zheng, Ya-Li ;
He, Liang-Tian ;
Pedrycz, Witold .
PROCEEDINGS OF 2016 INTERNATIONAL CONFERENCE ON WAVELET ANALYSIS AND PATTERN RECOGNITION (ICWAPR), 2016, :18-24
[23]   Robust Unsupervised Feature Selection by Nonnegative Sparse Subspace Learning [J].
Zheng, Wei ;
Yan, Hui ;
Yang, Jian ;
Yang, Jingyu .
2016 23RD INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2016, :3615-3620
[24]   Robust unsupervised feature selection by nonnegative sparse subspace learning [J].
Zheng, Wei ;
Yan, Hui ;
Yang, Jian .
NEUROCOMPUTING, 2019, 334 :156-171
[25]   Supervised feature selection by constituting a basis for the original space of features and matrix factorization [J].
Saberi-Movahed, Farid ;
Eftekhari, Mahdi ;
Mohtashami, Mohammad .
INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2020, 11 (07) :1405-1421
[26]   Partial multi-label feature selection via low-rank and sparse factorization with manifold learning [J].
Sun, Zhenzhen ;
Chen, Zexiang ;
Liu, Jinghua ;
Chen, Yewang ;
Yu, Yuanlong .
KNOWLEDGE-BASED SYSTEMS, 2024, 296
[27]   Subspace learning via Hessian regularized latent representation learning with l2,0-norm constraint: unsupervised feature selection [J].
Moslemi, Amir ;
Shaygani, Afshin .
INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024, 15 (11) :5361-5380
[28]   Subspace Clustering via Joint Unsupervised Feature Selection [J].
Dong, Wenhua ;
Wu, Xiao-Jun ;
Li, Hui ;
Feng, Zhen-Hua ;
Kittler, Josef .
2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, :3892-3898
[29]   RMFRASL: Robust Matrix Factorization with Robust Adaptive Structure Learning for Feature Selection [J].
Lai, Shumin ;
Huang, Longjun ;
Li, Ping ;
Luo, Zhenzhen ;
Wang, Jianzhong ;
Yi, Yugen .
ALGORITHMS, 2023, 16 (01)
[30]   Multi-label feature selection via feature manifold learning and sparsity regularization [J].
Cai, Zhiling ;
Zhu, William .
INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2018, 9 (08) :1321-1334