A comparative study of dimensionality reduction methods for large-scale image retrieval

被引:42
作者
Zhuo, Li [1 ]
Cheng, Bo [1 ]
Zhang, Jing [1 ]
机构
[1] Beijing Univ Technol, Signal & Informat Proc Lab, Beijing 100124, Peoples R China
基金
中国国家自然科学基金;
关键词
Large-scale image retrieval; Dimensionality reduction; OPTIMIZED SIFT; HSV histogram; Vocabulary tree;
D O I
10.1016/j.neucom.2014.03.014
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
"Curse of Dimensionality" is one of the important problems that Content-Based Image Retrieval (CBIR) confronts. Dimensionality reduction is an effective method to overcome it. In this paper, six commonly-used dimensionality reduction methods are compared and analyzed to examine their respective performance in image retrieval. The six methods include Principal Component Analysis (PCA), Fisher Linear Discriminant Analysis (FLDA), Local Fisher Discriminant Analysis (LFDA), Isometric Mapping (ISOMAP), Locally Linear Embedding (LLE), and Locality Preserving Projections (LPP). For comparison, Scale Invariant Feature Transform (SIFT) and color histogram in Hue, Saturation, Value (HSV) color space are firstly extracted as image features, meanwhile SIFT feature extraction procedure is optimized to reduce the number of SIFT features. Then, PCA, FLDA, LFDA, ISOMAP, LLE, and LPP are respectively applied to reduce the dimensions of feature vectors, which can be used to generate vocabulary trees. Finally, we can process large-scale image retrieval based on the inverted index built by vocabulary trees. In the experiments, the performance of various dimensionality reduction methods are analyzed comprehensively by comparing the retrieval performance, advantages and disadvantages, computational complexity and time-consuming of image retrieval. Through a series of experiments, we can conclude that dimensionality reduction method of LLE and LPP can effectively reduce computational complexity of image retrieval, while maintaining high retrieval performance. (C) 2014 Elsevier B.V. All rights reserved.
引用
收藏
页码:202 / 210
页数:9
相关论文
共 26 条
[1]  
Balasubramanian M., 2002, J SCI S5552, V295, pS9
[2]   Eigenfaces vs. Fisherfaces: Recognition using class specific linear projection [J].
Belhumeur, PN ;
Hespanha, JP ;
Kriegman, DJ .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 1997, 19 (07) :711-720
[3]   Laplacian eigenmaps for dimensionality reduction and data representation [J].
Belkin, M ;
Niyogi, P .
NEURAL COMPUTATION, 2003, 15 (06) :1373-1396
[4]  
Cheng B, 2014, PROCEEDINGS OF THE 2014 9TH INTERNATIONAL CONFERENCE ON COMPUTER VISION, THEORY AND APPLICATIONS (VISAPP 2014), VOL 2, P299
[5]  
Cormen T. H., 2009, INTRO ALGORITHMS
[6]  
Fanti C, 2004, ADV NEUR IN, V16, P1603
[7]   Visual-Textual Joint Relevance Learning for Tag-Based Social Image Search [J].
Gao, Yue ;
Wang, Meng ;
Zha, Zheng-Jun ;
Shen, Jialie ;
Li, Xuelong ;
Wu, Xindong .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2013, 22 (01) :363-376
[8]  
He X.F., 2004, LOCALITY PRESERVING, V2004153-160
[9]   Analysis of a complex of statistical variables into principal components [J].
Hotelling, H .
JOURNAL OF EDUCATIONAL PSYCHOLOGY, 1933, 24 :417-441
[10]   Nonlinear dimensionality reduction of data manifolds with essential loops [J].
Lee, JA ;
Verleysen, M .
NEUROCOMPUTING, 2005, 67 :29-53