Deep Joint-Semantics Reconstructing Hashing for Large-Scale Unsupervised Cross-Modal Retrieval

被引:254
作者
Su, Shupeng [1 ]
Zhong, Zhisheng [1 ]
Zhang, Chao [1 ]
机构
[1] Peking Univ, Sch EECS, Key Lab Machine Percept MOE, Beijing, Peoples R China
来源
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019) | 2019年
基金
中国国家自然科学基金; 国家重点研发计划;
关键词
D O I
10.1109/ICCV.2019.00312
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Cross-modal hashing encodes the multimedia data into a common binary hash space in which the correlations among the samples from different modalities can be effectively measured. Deep cross-modal hashing further improves the retrieval performance as the deep neural networks can generate more semantic relevant features and hash codes. In this paper, we study the unsupervised deep cross-modal hash coding and propose Deep Joint-Semantics Reconstructing Hashing (DJSRH), which has the following two main advantages. First, to learn binary codes that preserve the neighborhood structure of the original data, DJSRH constructs a novel joint-semantics affinity matrix which elaborately integrates the original neighborhood information from different modalities and accordingly is capable to capture the latent intrinsic semantic affinity for the input multi-modal instances. Second, DJSRH later trains the networks to generate binary codes that maximally reconstruct above joint-semantics relations via the proposed reconstructing framework, which is more competent for the batch-wise training as it reconstructs the specific similarity value unlike the common Laplacian constraint merely preserving the similarity order. Extensive experiments demonstrate the significant improvement by DJSRH in various cross-modal retrieval tasks.
引用
收藏
页码:3027 / 3035
页数:9
相关论文
共 50 条
[21]   Label guided correlation hashing for large-scale cross-modal retrieval [J].
Dong, Guohua ;
Zhang, Xiang ;
Lan, Long ;
Wang, Shiwei ;
Luo, Zhigang .
MULTIMEDIA TOOLS AND APPLICATIONS, 2019, 78 (21) :30895-30922
[22]   Joint-Modal Graph Convolutional Hashing for unsupervised cross-modal retrieval [J].
Meng, Hui ;
Zhang, Huaxiang ;
Liu, Li ;
Liu, Dongmei ;
Lu, Xu ;
Guo, Xinru .
NEUROCOMPUTING, 2024, 595
[23]   FDDH: Fast Discriminative Discrete Hashing for Large-Scale Cross-Modal Retrieval [J].
Liu, Xin ;
Wang, Xingzhi ;
Yiu-ming Cheung .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (11) :6306-6320
[24]   Semantic-consistent cross-modal hashing for large-scale image retrieval [J].
Gu, Xuesong ;
Dong, Guohua ;
Zhang, Xiang ;
Lan, Long ;
Luo, Zhigang .
NEUROCOMPUTING, 2021, 433 :181-198
[25]   Cross-Modal Self-Taught Hashing for large-scale image retrieval [J].
Xie, Liang ;
Zhu, Lei ;
Pan, Peng ;
Lu, Yansheng .
SIGNAL PROCESSING, 2016, 124 :81-92
[26]   Unsupervised Multi-modal Hashing for Cross-Modal Retrieval [J].
Yu, Jun ;
Wu, Xiao-Jun ;
Zhang, Donglin .
COGNITIVE COMPUTATION, 2022, 14 (03) :1159-1171
[27]   Unsupervised Multi-modal Hashing for Cross-Modal Retrieval [J].
Jun Yu ;
Xiao-Jun Wu ;
Donglin Zhang .
Cognitive Computation, 2022, 14 :1159-1171
[28]   Large-Scale Supervised Hashing for Cross-Modal Retreival [J].
Karbil, Loubna ;
Daoudi, Imane .
2017 IEEE/ACS 14TH INTERNATIONAL CONFERENCE ON COMPUTER SYSTEMS AND APPLICATIONS (AICCSA), 2017, :803-808
[29]   Unsupervised Deep Fusion Cross-modal Hashing [J].
Huang, Jiaming ;
Min, Chen ;
Jing, Liping .
ICMI'19: PROCEEDINGS OF THE 2019 INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, 2019, :358-366
[30]   Cross-Modal Semantic Embedding Hashing for Unsupervised Retrieval [J].
Zhang, Zhibin ;
Chen, Yun-fei .
2024 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN 2024, 2024,