Deep supervised fused similarity hashing for cross-modal retrieval

被引:0
|
作者
Ng W.W.Y. [1 ]
Xu Y. [1 ]
Tian X. [2 ]
Wang H. [3 ]
机构
[1] Guangdong Provincial Key Laboratory of Computational Intelligence and Cyberspace Information, School of Computer Science and Engineering, South China University of Technology, Guangzhou
[2] School of Artificial Intelligence, South China Normal University, Guangzhou
[3] School of Electronics, Electrical Engineering and Computer Science, Queens University Belfast, Belfast
基金
中国国家自然科学基金; 英国工程与自然科学研究理事会;
关键词
Common semantic subspace; Cross-modal retrieval; Deep hashing; Fused similarity;
D O I
10.1007/s11042-024-19581-2
中图分类号
学科分类号
摘要
The need for cross-modal retrieval increases significantly with the rapid growth of multimedia information on the Internet. However, most of existing cross-modal retrieval methods neglect the correlation between label similarity and intra-modality similarity in common semantic subspace training, which makes the trained common semantic subspace unable to preserve semantic similarity of original data effectively. Therefore, a novel cross-modal hashing method is proposed in this paper, namely, Deep Supervised Fused Similarity Hashing (DSFSH). The DSFSH mainly consists of two parts. Firstly, a fused similarity method is proposed to exploit the intrinsic inter-modality correlation of data while preserving the intra-modality relationship of data at the same time. Secondly, a novel quantization max-margin loss is proposed. The gap between cosine similarity and Hamming similarity is closed by minimizing this loss. Extensive experimental results on three benchmark datasets show that the proposed method yields better retrieval performance comparing to state-of-the-art methods. © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2024.
引用
收藏
页码:86537 / 86555
页数:18
相关论文
共 50 条
  • [41] Deep Adaptively-Enhanced Hashing With Discriminative Similarity Guidance for Unsupervised Cross-Modal Retrieval
    Shi, Yufeng
    Zhao, Yue
    Liu, Xin
    Zheng, Feng
    Ou, Weihua
    You, Xinge
    Peng, Qinmu
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (10) : 7255 - 7268
  • [42] Deep Feature-Based Neighbor Similarity Hashing With Adversarial Learning for Cross-Modal Retrieval
    Li, Kun
    Zhang, Yonghui
    Wang, Feng
    Liu, Guoxu
    Wei, Xianmin
    IEEE ACCESS, 2024, 12 : 128559 - 128569
  • [43] Deep Cross-Modal Hashing
    Jiang, Qing-Yuan
    Li, Wu-Jun
    30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 3270 - 3278
  • [44] Deep Unsupervised Momentum Contrastive Hashing for Cross-modal Retrieval
    Lu, Kangkang
    Yu, Yanhua
    Liang, Meiyu
    Zhang, Min
    Cao, Xiaowen
    Zhao, Zehua
    Yin, Mengran
    Xue, Zhe
    2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 126 - 131
  • [45] Deep semantic hashing with dual attention for cross-modal retrieval
    Jiagao Wu
    Weiwei Weng
    Junxia Fu
    Linfeng Liu
    Bin Hu
    Neural Computing and Applications, 2022, 34 : 5397 - 5416
  • [46] A triple fusion model for cross-modal deep hashing retrieval
    Hufei Wang
    Kaiqiang Zhao
    Dexin Zhao
    Multimedia Systems, 2023, 29 : 347 - 359
  • [47] Cross-Modal Hashing Retrieval Based on Deep Residual Network
    Li, Zhiyi
    Xu, Xiaomian
    Zhang, Du
    Zhang, Peng
    COMPUTER SYSTEMS SCIENCE AND ENGINEERING, 2021, 36 (02): : 383 - 405
  • [48] A triple fusion model for cross-modal deep hashing retrieval
    Wang, Hufei
    Zhao, Kaiqiang
    Zhao, Dexin
    MULTIMEDIA SYSTEMS, 2023, 29 (01) : 347 - 359
  • [49] A novel deep translated attention hashing for cross-modal retrieval
    Haibo Yu
    Ran Ma
    Min Su
    Ping An
    Kai Li
    Multimedia Tools and Applications, 2022, 81 : 26443 - 26461
  • [50] Semantic decomposition and enhancement hashing for deep cross-modal retrieval
    Fei, Lunke
    He, Zhihao
    Wong, Wai Keung
    Zhu, Qi
    Zhao, Shuping
    Wen, Jie
    PATTERN RECOGNITION, 2025, 160