Unsupervised feature selection guided by orthogonal representation of feature space

被引:27
作者
Jahani, Mahsa Samareh [1 ]
Aghamollaei, Gholamreza [1 ,2 ]
Eftekhari, Mahdi [3 ]
Saberi-Movahed, Farid [4 ]
机构
[1] Shahid Bahonar Univ Kerman, Fac Math & Comp, Dept Pure Math, Kerman, Iran
[2] Shahid Bahonar Univ Kerman, Mahani Math Res Ctr, Kerman, Iran
[3] Shahid Bahonar Univ Kerman, Dept Comp Engn, Kerman, Iran
[4] Grad Univ Adv Technol, Fac Sci & Modern Technol, Dept Appl Math, Kerman, Iran
关键词
Dimension reduction; Feature selection; Orthogonality; Correlation; Redundancy; INFORMATION;
D O I
10.1016/j.neucom.2022.10.030
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Feature selection has been an outstanding strategy in eliminating redundant and inefficient features in high-dimensional data. This paper introduces a novel unsupervised feature selection based on the matrix factorization, namely Unsupervised Feature Selection Guided by Orthogonal Representation (UFGOR). The orthogonality between a pair of variables refers to a specific case of linear independence such that they are perfectly uncorrelated. Motivated by the benefits of the orthogonality concept, the proposed UFGOR method is established based on the distance between the selected feature set and an orthogonal set corresponding to the whole feature space. Moreover, this orthogonal set is generated via QR-matrix factorization over the whole features and is employed as the compact representation of data matrix. In the next step, an unsupervised feature selection method is performed through the matrix factorization of the generated orthogonal set. Additionally, a dual-correlation model is utilized in the objective func-tion of UFGOR to simultaneously consider both the local correlation in a set of selected features and the global correlation among the samples of a data. A detailed convergence analysis in line with an effec-tive iterative algorithm proposed for the UFGOR method is also given. Numerical experiments on several real-world datasets illustrate the superior efficiency of our approach in comparison with some state-of-the-art unsupervised feature selection methods.(c) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页码:61 / 76
页数:16
相关论文
共 42 条
[1]   Manifold-respecting discriminant nonnegative matrix factorization [J].
An, Shounan ;
Yoo, Jiho ;
Choi, Seungjin .
PATTERN RECOGNITION LETTERS, 2011, 32 (06) :832-837
[2]  
[Anonymous], 2005, Adv. Neural Inf. Process. Syst.
[3]  
Cai D., 2010, Proceedings of the 16th ACM SIGKDD international conference on Knowledge discovery and data mining, P333, DOI DOI 10.1145/1835804.1835848
[4]   Graph Regularized Nonnegative Matrix Factorization for Data Representation [J].
Cai, Deng ;
He, Xiaofei ;
Han, Jiawei ;
Huang, Thomas S. .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2011, 33 (08) :1548-1560
[5]   Occam's razor in dimension reduction: Using reduced row Echelon form for finding linear independent features in high dimensional microarray datasets [J].
Ebrahimpour, Mohammad Kazem ;
Zare, Masoumeh ;
Eftekhari, Mahdi ;
Aghamolaei, Gholamreza .
ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2017, 62 :214-221
[6]  
Guyon I, 2006, STUD FUZZ SOFT COMP, V207, P1
[7]   Feature self-representation based hypergraph unsupervised feature selection via low-rank representation [J].
He, Wei ;
Cheng, Xiaohui ;
Hu, Rongyao ;
Zhu, Yonghua ;
Wen, Guoqiu .
NEUROCOMPUTING, 2017, 253 :127-134
[8]   Adaptive Component Embedding for Domain Adaptation [J].
Jing, Mengmeng ;
Zhao, Jidong ;
Li, Jingjing ;
Zhu, Lei ;
Yang, Yang ;
Shen, Heng Tao .
IEEE TRANSACTIONS ON CYBERNETICS, 2021, 51 (07) :3390-3403
[9]  
Jolliffe I.T., 1986, PRINCIPAL COMPONENT, P129, DOI 10.1007/978-1-4757-1904-8_8
[10]  
Lee DD, 2001, ADV NEUR IN, V13, P556