Robust unsupervised feature selection by nonnegative sparse subspace learning

被引:25
作者
Zheng, Wei [1 ,2 ,4 ,5 ]
Yan, Hui [1 ,3 ]
Yang, Jian [1 ,4 ,5 ]
机构
[1] Nanjing Univ Sci & Technol, Sch Comp Sci & Engn, Nanjing, Jiangsu, Peoples R China
[2] Jinling Inst Technol, Sch Comp Engn, Nanjing, Jiangsu, Peoples R China
[3] Nanjing Xiaozhuang Univ, Key Lab Trusted Cloud Comp & Big Data Anal, Nanjing, Jiangsu, Peoples R China
[4] Nanjing Univ Sci & Technol, Sch Comp Sci & Engn, Key Lab Intelligent Percept & Syst High Dimen, PCA Lab,Minist Educ, Nanjing, Jiangsu, Peoples R China
[5] Nanjing Univ Sci & Technol, Sch Comp Sci & Engn, Jiangsu Key Lab Image & Video Understanding Socia, Nanjing, Jiangsu, Peoples R China
关键词
Subspace learning; Non-negative matrix factorization; Unsupervised feature selection;
D O I
10.1016/j.neucom.2019.01.015
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Sparse subspace learning has been demonstrated to be effective in data mining and machine learning. In this paper, we cast the unsupervised feature selection scenario as a matrix factorization problem from the viewpoint of sparse subspace learning. By minimizing the reconstruction residual, the learned feature weight matrix with the l(2,1)-norm and the non-negative constraints not only removes the irrelevant features, but also captures the underlying low dimensional structure of the data points. Meanwhile in order to enhance the model's robustness, l(1) -norm error function is used to resistant to outliers and sparse noise. An efficient iterative algorithm is introduced to optimize this non-convex and non-smooth objective function and the proof of its convergence is given. Although, there is a subtraction item in our multiplicative update rule, we validate its non-negativity. The superiority of our model is demonstrated by comparative experiments on various original datasets with and without malicious pollution. (C) 2019 Published by Elsevier B.V.
引用
收藏
页码:156 / 171
页数:16
相关论文
共 41 条
[1]  
[Anonymous], 2007, COMPUTATIONAL METHOD
[2]  
[Anonymous], 2011, IJCAI INT JOINT C AR
[3]  
[Anonymous], 2009, ENCY COMPLEXITY SYST
[4]  
[Anonymous], P 20 ACM C INF KNOWL
[5]  
[Anonymous], 1951, P 2 BERK S
[6]  
Bezdek J. C., 2003, Neural, Parallel & Scientific Computations, V11, P351
[7]  
Cai D., 2010, P 16 ACM SIGKDD INT, P333, DOI [10.1145/1835804.1835848, DOI 10.1145/1835804.1835848]
[8]   Non-negative Matrix Factorization on Manifold [J].
Cai, Deng ;
He, Xiaofei ;
Wu, Xiaoyun ;
Han, Jiawei .
ICDM 2008: EIGHTH IEEE INTERNATIONAL CONFERENCE ON DATA MINING, PROCEEDINGS, 2008, :63-+
[9]   Robust Principal Component Analysis? [J].
Candes, Emmanuel J. ;
Li, Xiaodong ;
Ma, Yi ;
Wright, John .
JOURNAL OF THE ACM, 2011, 58 (03)
[10]  
Cover Thomas M, 2006, Elements of information theory