Subspace learning for feature selection via rank revealing QR factorization: Fast feature selection

被引:5
作者
Moslemi, Amir [1 ]
Ahmadian, Arash [2 ]
机构
[1] Seneca Polytech, Sch Software Design & Data Sci, Toronto, ON, Canada
[2] Univ Toronto, Edward S Rogers Sr Dept Elect & Comp Engn, Toronto, ON M5S 1A1, Canada
关键词
Feature selection; Rank revealing QR factorization; Non-negative matrix factorization; Genetic algorithm and hybrid feature selection; UNSUPERVISED FEATURE-SELECTION; SUPERVISED FEATURE-SELECTION; MATRIX FACTORIZATION; MUTUAL INFORMATION; CLASSIFICATION; OPTIMIZATION; ALGORITHMS; APPROXIMATION; REDUCTION; PATTERNS;
D O I
10.1016/j.eswa.2024.124919
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The identification of informative and distinguishing features from high-dimensional data has gained significant attention in the field of machine learning. Recently, there has been growing interest in employing matrix factorization-based techniques, such as non-negative matrix factorization, for feature selection. The primary objective of feature selection using matrix factorization is to extract a lower-dimensional subspace that captures the essence of the original space. This study introduces a novel unsupervised feature selection technique that leverages rank revealing QR (RRQR) factorization. Compared to singular value decomposition (SVD) and nonnegative matrix factorization (NMF), RRQR is more computationally efficient. The uniqueness of this technique lies in the utilization of the permutation matrix of QR for feature selection. Additionally, we integrate QR factorization into the objective function of NMF to create a new unsupervised feature selection method. Furthermore, we propose a hybrid feature selection algorithm by combining RRQR and a Genetic algorithm. The algorithm eliminates redundant features using RRQR factorization and selects the most distinguishing subset of features using the Genetic algorithm. Experimental comparisons with state-of-the-art feature selection algorithms in supervised, unsupervised, and semi-supervised settings demonstrate the reliability and robustness of the proposed algorithm. The evaluation is conducted on eight microarray datasets using KNN, SVM, and C4.5 classifiers. The experimental results indicate that the proposed method achieves comparable performance to the state-of-the-art feature selection methods. Our empirical findings demonstrate that the proposed method exhibits a significantly lower computational cost compared to other techniques.
引用
收藏
页数:18
相关论文
共 50 条
[31]   Feature selection by combining subspace learning with sparse representation [J].
Debo Cheng ;
Shichao Zhang ;
Xingyi Liu ;
Ke Sun ;
Ming Zong .
Multimedia Systems, 2017, 23 :285-291
[32]   Feature Selection for Learning-to-Rank using Simulated Annealing [J].
Allvi, Mustafa Wasif ;
Hasan, Mahamudul ;
Rayon, Lazim ;
Shahabuddin, Mohammad ;
Khan, Md Mosaddek ;
Ibrahim, Muhammad .
INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2020, 11 (03) :699-705
[33]   Feature selection in machine learning: A new perspective [J].
Cai, Jie ;
Luo, Jiawei ;
Wang, Shulin ;
Yang, Sheng .
NEUROCOMPUTING, 2018, 300 :70-79
[34]   Active learning of constraints for weighted feature selection [J].
Hijazi, Samah ;
Hamad, Denis ;
Kalakech, Mariam ;
Kalakech, Ali .
ADVANCES IN DATA ANALYSIS AND CLASSIFICATION, 2021, 15 (02) :337-377
[35]   Unsupervised feature selection via local structure learning and sparse learning [J].
Lei, Cong ;
Zhu, Xiaofeng .
MULTIMEDIA TOOLS AND APPLICATIONS, 2018, 77 (22) :29605-29622
[36]   Feature Subspace Learning-Based Binary Differential Evolution Algorithm for Unsupervised Feature Selection [J].
Li, Tao ;
Qian, Yuhua ;
Li, Feijiang ;
Liang, Xinyan ;
Zhan, Zhi-Hui .
IEEE TRANSACTIONS ON BIG DATA, 2025, 11 (01) :99-114
[37]   Joint subspace learning and subspace clustering based unsupervised feature selection [J].
Xiao, Zijian ;
Chen, Hongmei ;
Mi, Yong ;
Luo, Chuan ;
Horng, Shi-Jinn ;
Li, Tianrui .
NEUROCOMPUTING, 2025, 635
[38]   Feature selection in machine learning via variable neighborhood search [J].
Syed, Mujahid N. .
OPTIMIZATION LETTERS, 2023, 17 (09) :2321-2345
[39]   A novel feature selection method and its application [J].
Li, Bing ;
Chow, Tommy W. S. ;
Huang, Di .
JOURNAL OF INTELLIGENT INFORMATION SYSTEMS, 2013, 41 (02) :235-268
[40]   Fast Multilabel Feature Selection via Global Relevance and Redundancy Optimization [J].
Zhang, Jia ;
Lin, Yidong ;
Jiang, Min ;
Li, Shaozi ;
Tang, Yong ;
Long, Jinyi ;
Weng, Jian ;
Tan, Kay Chen .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (04) :5721-5734