Unsupervised Deep Fusion Cross-modal Hashing

被引:4
|
作者
Huang, Jiaming [1 ]
Min, Chen [1 ]
Jing, Liping [1 ]
机构
[1] Beijing Jiaotong Univ, Beijing, Peoples R China
基金
北京市自然科学基金; 中国国家自然科学基金;
关键词
multimodal; information retrieval; hashing; deep learning;
D O I
10.1145/3340555.3353752
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
To handle the large-scale data in terms of storage and searching time, learning to hash becomes popular due to its efficiency and effectiveness in approximate cross-modal nearest neighbors searching. Most existing unsupervised cross modal hashing methods, to shorten the semantic gap, try to simultaneously minimize the loss of intra-modal similarity and the loss of inter-modal similarity. However, these models can not guarantee in theory these two losses are simultaneously minimized. In this paper, we first theoretically proved that cross-modal hashing could be implemented by protecting both intra-modal and inter-modal similarity with the aid of variational inference technique and point out the problem that maximizing intra and inter-modal similarity is mutually constrained. In this case, we propose an unsupervised cross-modal hashing framework named as Unsupervised Deep Fusion Cross-modal Hashing (UDFCH) which leverages the data fusion to capture the underlying manifold across modalities to avoid above problem. What's more, in order to reduce the quantization loss, we sample hash codes from different Bernoulli distributions through a reparameterization trick. Our UDFCH framework has two stages. The first stage aims at mining the the intra-modal structure of each modality. The second stage aims to determine the modality-aware hash code by sufficiently considering the correlation and manifold structure among modalities. A series of experiments conducted on three benchmark datasets show that the proposed UDFCH framework outperforms the state-of-the-art methods on different cross-modal retrieval tasks.
引用
收藏
页码:358 / 366
页数:9
相关论文
共 50 条
  • [1] Deep Unsupervised Momentum Contrastive Hashing for Cross-modal Retrieval
    Lu, Kangkang
    Yu, Yanhua
    Liang, Meiyu
    Zhang, Min
    Cao, Xiaowen
    Zhao, Zehua
    Yin, Mengran
    Xue, Zhe
    2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 126 - 131
  • [2] Unsupervised Contrastive Cross-Modal Hashing
    Hu, Peng
    Zhu, Hongyuan
    Lin, Jie
    Peng, Dezhong
    Zhao, Yin-Ping
    Peng, Xi
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (03) : 3877 - 3889
  • [3] Deep Consistency Preserving Network for Unsupervised Cross-Modal Hashing
    Li, Mengluan
    Guo, Yanqing
    Fu, Haiyan
    Li, Yi
    Su, Hong
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT I, 2024, 14425 : 235 - 246
  • [4] Unsupervised Deep Imputed Hashing for Partial Cross-modal Retrieval
    Chen, Dong
    Cheng, Miaomiao
    Min, Chen
    Jing, Liping
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [5] FUSION-SUPERVISED DEEP CROSS-MODAL HASHING
    Wang, Li
    Zhu, Lei
    Yu, En
    Sun, Jiande
    Zhang, Huaxiang
    2019 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2019, : 37 - 42
  • [6] Deep Multiscale Fusion Hashing for Cross-Modal Retrieval
    Nie, Xiushan
    Wang, Bowei
    Li, Jiajia
    Hao, Fanchang
    Jian, Muwei
    Yin, Yilong
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2021, 31 (01) : 401 - 410
  • [7] Completely Unsupervised Cross-Modal Hashing
    Duan, Jiasheng
    Zhang, Pengfei
    Huang, Zi
    DATABASE SYSTEMS FOR ADVANCED APPLICATIONS (DASFAA 2020), PT I, 2020, 12112 : 178 - 194
  • [8] Deep Cross-Modal Hashing
    Jiang, Qing-Yuan
    Li, Wu-Jun
    30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 3270 - 3278
  • [9] Unsupervised Deep Cross-Modal Hashing by Knowledge Distillation for Large-scale Cross-modal Retrieval
    Li, Mingyong
    Wang, Hongya
    PROCEEDINGS OF THE 2021 INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL (ICMR '21), 2021, : 183 - 191
  • [10] A triple fusion model for cross-modal deep hashing retrieval
    Hufei Wang
    Kaiqiang Zhao
    Dexin Zhao
    Multimedia Systems, 2023, 29 : 347 - 359