Work Together: Correlation-Identity Reconstruction Hashing for Unsupervised Cross-Modal Retrieval

被引:61
作者
Zhu, Lei [1 ]
Wu, Xize [1 ]
Li, Jingjing [2 ]
Zhang, Zheng [3 ]
Guan, Weili [4 ]
Shen, Heng Tao [2 ]
机构
[1] Shandong Normal Univ, Sch Informat Sci & Engn, Jinan 250358, Peoples R China
[2] Univ Elect Sci & Technol China, Sch Comp Sci & Engn, Chengdu 611731, Peoples R China
[3] Harbin Inst Technol, Shenzhen Key Lab Visual Object Detect & Recognit, Shenzhen 518055, Peoples R China
[4] Monash Univ, Fac Informat Technol, Clayton Campus, Clayton, Vic 3800, Australia
基金
中国国家自然科学基金;
关键词
Cross-modal retrieval; correlation-identity; reconstruction network; multi-modal correlation; semantic reconstruction; NETWORK;
D O I
10.1109/TKDE.2022.3218656
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Unsupervised cross-modal hashing has attracted considerable attention to support large-scale cross-modal retrieval. Although promising progresses have been made so far, existing methods still suffer from limited capability on excavating and preserving the intrinsic multi-modal semantics. In this paper, we propose a Correlation-Identity Reconstruction Hashing (CIRH) method to alleviate this challenging problem. We develop a new unsupervised deep cross-modal hash learning framework to model and preserve the heterogeneous multi-modal correlation semantics into both hash codes and functions, and simultaneously, we involve both the hash codes and functions with the descriptive identity semantics. Specifically, we construct a multi-modal collaborated graph to model the heterogeneous multi-modal correlations, and jointly perform the intra-modal and cross-modal semantic aggregation on homogeneous and heterogeneous graph networks to generate a multi-modal complementary representation with correlation reconstruction. Furthermore, an identity semantic reconstruction process is designed to involve the generated representation with identity semantics by reconstructing the input modality representations. Finally, we propose a correlation-identity consistent hash function learning strategy to transfer the modelled multi-modal semantics into the neural networks of modality-specific deep hash functions. Experiments demonstrate the superior performance of the proposed method on both retrieval accuracy and efficiency. We provide our source codes and experimental datasets at https://github.com/XizeWu/CIRH
引用
收藏
页码:8838 / 8851
页数:14
相关论文
共 59 条
[11]   Unsupervised Deep Cross-modality Spectral Hashing [J].
Hoang, Tuan ;
Do, Thanh-Toan ;
Nguyen, Tam V. ;
Cheung, Ngai-Man .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 :8391-8406
[12]   Deep Binary Reconstruction for Cross-Modal Hashing [J].
Hu, Di ;
Nie, Feiping ;
Li, Xuelong .
IEEE TRANSACTIONS ON MULTIMEDIA, 2019, 21 (04) :973-985
[13]   Creating Something from Nothing: Unsupervised Knowledge Distillation for Cross-Modal Hashing [J].
Hu, Hengtong ;
Xie, Lingxi ;
Hong, Richang ;
Tian, Qi .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, :3120-3129
[14]  
Huiskes M., 2008, Proc. Multimedia Infor- mation Retrieval, P39
[15]   Deep Cross-Modal Hashing [J].
Jiang, Qing-Yuan ;
Li, Wu-Jun .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :3270-3278
[16]  
Kingsbury D, 2015, P1, DOI [10.48550/arXiv.1312.6114, DOI 10.48550/ARXIV.1412.6980, 10.48550/arXiv.1412.6980, DOI 10.1021/bk-2015-1214.ch001]
[17]  
Ko Y, 2012, SIGIR 2012: PROCEEDINGS OF THE 35TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, P1029
[18]   ImageNet Classification with Deep Convolutional Neural Networks [J].
Krizhevsky, Alex ;
Sutskever, Ilya ;
Hinton, Geoffrey E. .
COMMUNICATIONS OF THE ACM, 2017, 60 (06) :84-90
[19]  
Kumar Shaishav., 2011, International Joint Conference on Artificial Intelligence (IJCAI), P1360, DOI DOI 10.5591/978-1-57735-516-8/IJCAI11-23
[20]   Adaptive Label Correlation Based Asymmetric Discrete Hashing for Cross-Modal Retrieval [J].
Li, Huaxiong ;
Zhang, Chao ;
Jia, Xiuyi ;
Gao, Yang ;
Chen, Chunlin .
IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (02) :1185-1199