Unsupervised feature selection via adaptive hypergraph regularized latent representation learning

被引:22
作者
Ding, Deqiong [1 ]
Yang, Xiaogao [2 ]
Xia, Fei [3 ]
Ma, Tiefeng [1 ]
Liu, Haiyun [4 ]
Tang, Chang [5 ]
机构
[1] Southwestern Univ Finance & Econ, Sch Stat, Chengdu 611130, Sichuan, Peoples R China
[2] Hunan Univ Arts & Sci, Coll Mech Engn, Changde 415000, Peoples R China
[3] Naval Univ Engn, NUDT, Opt Engn Postdoctoral Mobile Stn, Changsha 410073, Peoples R China
[4] Xuzhou Med Univ, Inst Cardiovasc Dis Res, Affiliated Huaian Hosp, Huaian 223100, Peoples R China
[5] China Univ Geosci, Sch Comp Sci, Wuhan 430074, Peoples R China
基金
中国博士后科学基金; 中国国家自然科学基金;
关键词
Unsupervised feature selection; Hypergraph learning; Latent representation learning; Local structure preservation; STRUCTURE PRESERVATION; GRAPH;
D O I
10.1016/j.neucom.2019.10.018
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Due to the rapid development of multimedia technology, a large number of unlabelled data with high dimensionality need to be processed. The high dimensionality of data not only increases the computation burden of computer hardware, but also hinders algorithms to obtain optimal performance. Unsupervised feature selection, which is regarded as a means of dimensionality reduction, has been widely recognized as an important and challenging pre-step for many machine learning and data mining tasks. However, we observe that there are at least two issues in previous unsupervised feature selection methods. Firstly, traditional unsupervised feature selection algorithms usually assume that the data instances are identically distributed and there is no dependency between them. However, the data instances are not only associated with high dimensional features but also inherently interconnected with each other. Secondly, the traditional similarity graph used in previous methods can only describe the pair-wise relations of data, but cannot capture the high-order relations, so that the complex structures implied in the data cannot be sufficiently exploited. In this work, we propose a robust unsupervised feature selection method which embeds the latent representation learning into feature selection. Instead of measuring the feature importances in original data space, the feature selection is carried out in the learned latent representation space which is more robust to noises. In order to capture the local manifold geometrical structure of original data in a high-order manner, a hypergraph is adaptively learned and embedded into the resultant model. An efficient alternating algorithm is developed to optimize the problem. Experimental results on eight benchmark data sets demonstrate the effectiveness of the proposed method. (C) 2019 Elsevier B.V. All rights reserved.
引用
收藏
页码:79 / 97
页数:19
相关论文
共 81 条
[1]   The Effect of the Characteristics of the Dataset on the Selection Stability [J].
Alelyani, Salem ;
Liu, Huan ;
Wang, Lei .
2011 23RD IEEE INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2011), 2011, :970-977
[2]  
[Anonymous], P INT C MACH LEARN
[3]  
[Anonymous], 2006, P ADV NEUR INF PROC
[4]  
[Anonymous], 2014, P SIAM INT C DAT MIN
[5]   Deterministic Feature Selection for k-Means Clustering [J].
Boutsidis, Christos ;
Magdon-Ismail, Malik .
IEEE TRANSACTIONS ON INFORMATION THEORY, 2013, 59 (09) :6099-6110
[6]  
Boutsidis P., 2009, Advances in Neural Information Processing Systems, P153
[7]  
Cai D., 2010, P 16 ACM SIGKDD INT, P333
[8]   Semi-supervised graph-based hyperspectral image classification [J].
Camps-Valls, Gustavo ;
Bandos, Tatyana V. ;
Zhou, Dengyong .
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2007, 45 (10) :3044-3054
[9]   A biological mechanism for Bayesian feature selection: Weight decay and raising the LASSO [J].
Connor, Patrick ;
Hollensen, Paul ;
Krigolson, Olav ;
Trappenberg, Thomas .
NEURAL NETWORKS, 2015, 67 :121-130
[10]   Bayesian feature and model selection for Gaussian mixture models [J].
Constantinopoulos, C ;
Titsias, MK ;
Likas, A .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2006, 28 (06) :1013-U1