Dual-graph with non-convex sparse regularization for multi-label feature selection

被引:6
作者
Sun, Zhenzhen [1 ,2 ]
Xie, Hao [1 ]
Liu, Jinghua [1 ]
Gou, Jin [1 ]
Yu, Yuanlong [3 ]
机构
[1] HuaQiao Univ, Coll Comp Sci & Technol, Jimei Ave, Xiamen 361021, Fujian, Peoples R China
[2] HuaQiao Univ, Xiamen Key Lab Comp Vis & Pattern Recognit, Jimei Ave, Xiamen 361021, Fujian, Peoples R China
[3] Fuzhou Univ, Coll Comp & Data Sci, Wulong Jiangbei Ave, Fuzhou 350108, Fujian, Peoples R China
基金
中国国家自然科学基金;
关键词
Feature selection; Multi-label learning; Dual graph; Non-convex sparse regularization; MUTUAL INFORMATION; LAPLACIAN SCORE;
D O I
10.1007/s10489-023-04515-6
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
As for single-label learning, feature selection has become a crucial data pre-processing tool for multi-label learning due to its ability to reduce the feature dimensionality to mitigate the influence of "the curse of dimensionality". In recent years, the multi-label feature selection (MFS) methods based on manifold learning and sparse regression have confirmed their superiority and gained more and more attention than other methods. However, most of these methods only consider exploiting the geometric structure distributed in the data manifold but ignore the geometric structure distributed in the feature manifold, resulting in incomplete information about the learned manifold. Aiming to simultaneously exploit the geometric information of data and feature spaces for feature selection, this paper proposes a novel MFS method named dual-graph based multi-label feature selection (DGMFS). In the framework of DGMFS, two nearest neighbor graphs are constructed to form a dual-graph regularization to explore the geometric structures of both the data manifold and the feature manifold simultaneously. Then, we combined the dual-graph regularization with a non-convex sparse regularization l(2,1 - 2)-norm to obtain a more sparse solution for the weight matrix, which can better deal with the redundancy features. Finally, we adopt the EM strategy to design an iterative updating optimization algorithm for DGMFS and provide the convergence analysis of the optimization algorithm. Extensive experimental results on ten benchmark multi-label data sets have shown that DGMFS is more effective than several state-of-the-art MFS methods.
引用
收藏
页码:21227 / 21247
页数:21
相关论文
共 41 条
[1]   Soft-constrained Laplacian score for semi-supervised multi-label feature selection [J].
Alalga, Abdelouahid ;
Benabdeslem, Khalid ;
Taleb, Nora .
KNOWLEDGE AND INFORMATION SYSTEMS, 2016, 47 (01) :75-98
[2]   Multi-Label Feature Selection using Correlation Information [J].
Braytee, Ali ;
Liu, Wei ;
Catchpoole, Daniel R. ;
Kennedy, Paul J. .
CIKM'17: PROCEEDINGS OF THE 2017 ACM CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, 2017, :1649-1656
[3]  
Cai D., 2010, P 16 ACM SIGKDD INT, P333
[4]   Multi-label feature selection via feature manifold learning and sparsity regularization [J].
Cai, Zhiling ;
Zhu, William .
INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2018, 9 (08) :1321-1334
[5]  
Demsar J, 2006, J MACH LEARN RES, V7, P1
[6]   Multiple graph unsupervised feature selection [J].
Du, Xingzhong ;
Yan, Yan ;
Pan, Pingbo ;
Long, Guodong ;
Zhao, Lei .
SIGNAL PROCESSING, 2016, 120 :754-760
[7]   Manifold learning with structured subspace for multi-label feature selection [J].
Fan, Yuling ;
Liu, Jinghua ;
Liu, Peizhong ;
Du, Yongzhao ;
Lan, Weiyao ;
Wu, Shunxiang .
PATTERN RECOGNITION, 2021, 120
[8]   Multi-label feature selection with local discriminant model and label correlations [J].
Fan, Yuling ;
Liu, Jinghua ;
Weng, Wei ;
Chen, Baihua ;
Chen, Yannan ;
Wu, Shunxiang .
NEUROCOMPUTING, 2021, 442 :98-115
[9]  
Gu QQ, 2009, KDD-09: 15TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, P359
[10]   An efficient Pareto-based feature selection algorithm for multi-label classification [J].
Hashemi, Amin ;
Dowlatshahi, Mohammad Bagher ;
Nezamabadi-pour, Hossein .
INFORMATION SCIENCES, 2021, 581 :428-447