Deep Self-Taught Hashing for Image Retrieval

被引:38
作者
Liu, Yu [1 ]
Song, Jingkuan [2 ,3 ]
Zhou, Ke [1 ]
Yan, Lingyu [4 ]
Liu, Li [5 ]
Zou, Fuhao [6 ]
Shao, Ling [5 ]
机构
[1] Huazhong Univ Sci & Technol, Wuhan Natl Lab Optoelect, Key Lab Informat Storage Syst, Sch Comp Sci & Technol,Minist Educ China, Wuhan 430074, Hubei, Peoples R China
[2] Univ Elect Sci & Technol China, Ctr Future Media, Chengdu 611731, Sichuan, Peoples R China
[3] Univ Elect Sci & Technol China, Sch Comp Sci, Chengdu 611731, Sichuan, Peoples R China
[4] Hubei Univ Technol, Sch Comp Sci, Wuhan 430068, Hubei, Peoples R China
[5] Univ East Anglia, Sch Comp Sci, Norwich NR1 1NN, Norfolk, England
[6] Huazhong Univ Sci & Technol, Sch Comp Sci & Technol, Wuhan 430074, Hubei, Peoples R China
基金
中国国家自然科学基金;
关键词
Deep Learning; hashing; image retrieval; self-taught; QUANTIZATION;
D O I
10.1109/TCYB.2018.2822781
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Hashing algorithm has been widely used to speed up image retrieval due to its compact binary code and fast distance calculation. The combination with deep learning boosts the performance of hashing by learning accurate representations and complicated hashing functions. So far, the most striking success in deep hashing have mostly involved discriminative models, which require labels. To apply deep hashing on datasets without labels, we propose a deep self-taught hashing algorithm (DSTH), which generates a set of pseudo labels by analyzing the data itself, and then learns the hash functions for novel data using discriminative deep models. Furthermore, we generalize DSTH to support both supervised and unsupervised cases by adaptively incorporating label information. We use two different deep learning framework to train the hash functions to deal with out-of-sample problem and reduce the time complexity without loss of accuracy. We have conducted extensive experiments to investigate different settings of DSTH, and compared it with state-of-the-art counterparts in six publicly available datasets. The experimental results show that DSTH outperforms the others in all datasets.
引用
收藏
页码:2229 / 2241
页数:13
相关论文
共 46 条
[1]  
[Anonymous], 2018, P AAAI NEW ORL LA US
[2]  
[Anonymous], 2018, P AAAI
[3]  
[Anonymous], 2009, NIPS
[4]  
[Anonymous], 2015, PROC CVPR IEEE
[5]   Total recall: Automatic query expansion with a generative feature model for object retrieval [J].
Chum, Ondrej ;
Philbin, James ;
Sivic, Josef ;
Isard, Michael ;
Zisserman, Andrew .
2007 IEEE 11TH INTERNATIONAL CONFERENCE ON COMPUTER VISION, VOLS 1-6, 2007, :496-+
[6]   Histograms of oriented gradients for human detection [J].
Dalal, N ;
Triggs, B .
2005 IEEE COMPUTER SOCIETY CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, VOL 1, PROCEEDINGS, 2005, :886-893
[7]  
Donahue J, 2014, PR MACH LEARN RES, V32
[8]   Video Captioning With Attention-Based LSTM and Semantic Consistency [J].
Gao, Lianli ;
Guo, Zhao ;
Zhang, Hanwang ;
Xu, Xing ;
Shen, Heng Tao .
IEEE TRANSACTIONS ON MULTIMEDIA, 2017, 19 (09) :2045-2055
[9]   Scalable Multimedia Retrieval by Deep Learning Hashing with Relative Similarity Learning [J].
Gao, Lianli ;
Song, Jingkuan ;
Zou, Fuhao ;
Zhang, Dongxiang ;
Shao, Jie .
MM'15: PROCEEDINGS OF THE 2015 ACM MULTIMEDIA CONFERENCE, 2015, :903-906
[10]   Learning in high-dimensional multimedia data: the state of the art [J].
Gao, Lianli ;
Song, Jingkuan ;
Liu, Xingyi ;
Shao, Junming ;
Liu, Jiajun ;
Shao, Jie .
MULTIMEDIA SYSTEMS, 2017, 23 (03) :303-313