A Dictionary-Based Algorithm for Dimensionality Reduction and Data Reconstruction

被引:1
作者
Zhao, Zhong [1 ]
Feng, Guocan [1 ]
机构
[1] Sun Yat Sen Univ, Sch Math & Computat Sci, Guangzhou 510275, Guangdong, Peoples R China
来源
2014 22ND INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR) | 2014年
关键词
Dimensionality reduction; data reconstruction; dictionary learning;
D O I
10.1109/ICPR.2014.276
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Nonlinear dimensionality reduction (DR) is a basic problem in manifold learning. However, many DR algorithms cannot deal with the out-of-sample extension problem and thus cannot be used in large-scale DR problem. Furthermore, many DR algorithms only consider how to reduce the dimensionality but seldom involve with how to reconstruct the original high dimensional data from the low dimensional embeddings (i.e. data reconstruction problem). In this paper, we propose a dictionary-based algorithm to deal with the out-of-sample extension problem for large-scale DR task. In this algorithm, we train a high dimensional dictionary and a low dimensional dictionary corresponding to the high dimensional data and their low dimensional embeddings respectively. With these two dictionaries, dimensionality reduction and data reconstruction can be easily conducted by coding the input data point over one dictionary, and then use the code to predict the output data point over another dictionary. Compared to the existing DR algorithms, our algorithm has high efficiency since analytic solution is derived. Besides, our reconstruction algorithm can be applied to many DR algorithms to make them have the ability to perform data reconstruction. Experiments on synthetic datasets and real world datasets show that, for both dimensionality reduction and data reconstruction, our algorithm is accurate and fast.
引用
收藏
页码:1556 / 1561
页数:6
相关论文
共 22 条
[1]  
[Anonymous], COMP VIS ICCV 2011 I
[2]  
[Anonymous], INT S SIGN PROC ITS
[3]  
Belkin M, 2002, ADV NEUR IN, V14, P585
[4]  
Boothby William M., 1986, Pure and Applied Mathematics, V120
[5]  
Cox Michael AA Cox TrevorF., 2010, Multidimensional scaling
[6]   Hessian eigenmaps: Locally linear embedding techniques for high-dimensional data [J].
Donoho, DL ;
Grimes, C .
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2003, 100 (10) :5591-5596
[7]  
He XF, 2004, ADV NEUR IN, V16, P153
[8]   Reducing the dimensionality of data with neural networks [J].
Hinton, G. E. ;
Salakhutdinov, R. R. .
SCIENCE, 2006, 313 (5786) :504-507
[9]   Incremental Laplacian eigenmaps by preserving adjacent information between data points [J].
Jia, Peng ;
Yin, Junsong ;
Huang, Xinsheng ;
Hu, Dewen .
PATTERN RECOGNITION LETTERS, 2009, 30 (16) :1457-1463
[10]  
Jolliffe I., 2002, PRINCIPAL COMPONENT, DOI [10.1007/978-1-4757-1904-8_7, 10.1016/0169-7439(87)80084-9]