Automatic Subspace Learning via Principal Coefficients Embedding

被引:78
作者
Peng, Xi [1 ]
Lu, Jiwen [2 ]
Yi, Zhang [3 ]
Yan, Rui [3 ]
机构
[1] ASTAR, Inst Infocomm Res, Singapore 138632, Singapore
[2] Tsinghua Univ, Dept Automat, Beijing 100084, Peoples R China
[3] Sichuan Univ, Coll Comp Sci, Chengdu 610065, Sichuan, Peoples R China
基金
中国国家自然科学基金;
关键词
Automatic dimension reduction; corrupted data; graph embedding; manifold learning; robustness; FACE RECOGNITION; INTRINSIC DIMENSION; FRAMEWORK; REPRESENTATION;
D O I
10.1109/TCYB.2016.2572306
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, we address two challenging problems in unsupervised subspace learning: 1) how to automatically identify the feature dimension of the learned subspace (i.e., automatic subspace learning) and 2) how to learn the underlying subspace in the presence of Gaussian noise (i.e., robust subspace learning). We show that these two problems can be simultaneously solved by proposing a new method [(called principal coefficients embedding (PCE)]. For a given data set D is an element of R-mxn, PCE recovers a clean data set D-0 is an element of R-mxn from D and simultaneously learns a global reconstruction relation C is an element of R-nxn of D-0. By preserving C into an m'-dimensional space, the proposed method obtains a projection matrix that can capture the latent manifold structure of D-0, where m' << m is automatically determined by the rank of C with theoretical guarantees. PCE has three advantages: 1) it can automatically determine the feature dimension even though data are sampled from a union of multiple linear subspaces in presence of the Gaussian noise; 2) although the objective function of PCE only considers the Gaussian noise, experimental results show that it is robust to the non-Gaussian noise (e.g., random pixel corruption) and real disguises; and 3) our method has a closed-form solution and can be calculated very fast. Extensive experimental results show the superiority of PCE on a range of databases with respect to the classification accuracy, robustness, and efficiency.
引用
收藏
页码:3583 / 3596
页数:14
相关论文
共 78 条
[1]  
[Anonymous], 2004, NIPS
[2]  
[Anonymous], 1996, document CUCS-006-96
[3]  
[Anonymous], 2007, PROC IEEE INT C COMP
[4]  
[Anonymous], 2009, P INT WORKSH COMP AD
[5]  
[Anonymous], 1998, 24 CVC
[6]  
[Anonymous], 2010, UCBEECS201013
[7]   General Subspace Learning With Corrupted Training Data Via Graph Embedding [J].
Bao, Bing-Kun ;
Liu, Guangcan ;
Hong, Richang ;
Yan, Shuicheng ;
Xu, Changsheng .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2013, 22 (11) :4380-4393
[8]   Fast low-rank modifications of the thin singular value decomposition [J].
Brand, M .
LINEAR ALGEBRA AND ITS APPLICATIONS, 2006, 415 (01) :20-30
[9]   Estimating the intrinsic dimension of data with a fractal-based method [J].
Camastra, F ;
Vinciarelli, A .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2002, 24 (10) :1404-1407
[10]   Robust Principal Component Analysis? [J].
Candes, Emmanuel J. ;
Li, Xiaodong ;
Ma, Yi ;
Wright, John .
JOURNAL OF THE ACM, 2011, 58 (03)