Precise feature selection via non-convex regularized graph embedding and self-representation for unsupervised learning

被引:1
作者
Bai, Hanru [1 ]
Huang, Ming [1 ]
Zhong, Ping [1 ]
机构
[1] China Agr Univ, Coll Sci, Beijing 100083, Peoples R China
关键词
Unsupervised feature selection; Self-representation; Graph optimization; e'20-norm constraint; ALGORITHM;
D O I
10.1016/j.knosys.2024.111900
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In graph embedding learning with unsupervised feature selection fields, similarity matrices are usually obtained only from the initial noise -laden samples, and the probable association between various features is often ignored. To tackle the issues described above, this article proposes an effective unsupervised group feature selection approach via non -convex regularized graph embedding and self -representation (NLGMS). NLGMS integrates the similarity matrix and feature selection as a joint system on the basis of graph. It combines selfrepresentation learning into the system to preserve the global structure and learns the nonconvex projection matrix to direct the procedure of feature selection. Specifically, first, NLGMS presents a unified framework for global structure learning, local structure learning, and feature selection process so the data's intrinsic structure is excellently captured and learned adaptively. Second, nonconvex e' 2 , 0 -norm is enforced upon the projection matrix to perform the process of feature selection, ensuring that the optimal feature subset can be picked precisely. Nonconvex e' 2 , 0 -norm constraint considers the probable association between various features so that it can deliver optimal performance by considering the feature subset as an integral whole, which other convex regularized constraints often fail to achieve. Finally, the introduction to self -representation learning can facilitate the reconstruction of samples in low -rank subspaces, strengthening the model's robustness. To tackle this challenging nonconvex model, an innovative alternative optimization technique is exploited. In addition, to validate the superiority of NLGMS, exhaustive experiments are performed on one synthetic dataset, six benchmark datasets, and one fish image dataset. The source code is available at: https://github.com/hrbai/ NLGMS.
引用
收藏
页数:15
相关论文
共 54 条
[1]  
[Anonymous], 2013, Int. Joint Conf. Artif. Intell
[2]  
Artacho F., 2019, Convex Optimization, Vthird
[3]  
Benavoli A, 2017, J MACH LEARN RES, V18
[4]  
Cai D., 2010, P 16 ACM SIGKDD INT, P333, DOI DOI 10.1145/1835804.1835848
[5]   A SINGULAR VALUE THRESHOLDING ALGORITHM FOR MATRIX COMPLETION [J].
Cai, Jian-Feng ;
Candes, Emmanuel J. ;
Shen, Zuowei .
SIAM JOURNAL ON OPTIMIZATION, 2010, 20 (04) :1956-1982
[6]   Unsupervised Feature Selection with Adaptive Structure Learning [J].
Du, Liang ;
Shen, Yi-Dong .
KDD'15: PROCEEDINGS OF THE 21ST ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2015, :209-218
[7]   Exploiting Combination Effect for Unsupervised Feature Selection by l2,0 Norm [J].
Du, Xingzhong ;
Nie, Feiping ;
Wang, Weiqing ;
Yang, Yi ;
Zhou, Xiaofang .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2019, 30 (01) :201-214
[8]   ON THE DOUGLAS-RACHFORD SPLITTING METHOD AND THE PROXIMAL POINT ALGORITHM FOR MAXIMAL MONOTONE-OPERATORS [J].
ECKSTEIN, J ;
BERTSEKAS, DP .
MATHEMATICAL PROGRAMMING, 1992, 55 (03) :293-318
[9]   Manifold learning with structured subspace for multi-label feature selection [J].
Fan, Yuling ;
Liu, Jinghua ;
Liu, Peizhong ;
Du, Yongzhao ;
Lan, Weiyao ;
Wu, Shunxiang .
PATTERN RECOGNITION, 2021, 120
[10]   A unified low-order information-theoretic feature selection framework for multi-label learning [J].
Gao, Wanfu ;
Hao, Pingting ;
Wu, Yang ;
Zhang, Ping .
PATTERN RECOGNITION, 2023, 134