Cross-modal hashing retrieval with compatible triplet representation

被引:0
|
作者
Hao, Zhifeng [1 ]
Jin, Yaochu [2 ]
Yan, Xueming [1 ,3 ]
Wang, Chuyue [3 ]
Yang, Shangshang [4 ]
Ge, Hong [5 ]
机构
[1] Shantou Univ, Key Lab Intelligent Mfg Technol, Shantou 515063, Guangdong, Peoples R China
[2] Westlake Univ, Sch Engn, Hangzhou 310030, Peoples R China
[3] Guangdong Univ Foreign Studies, Sch Informat Sci & Technol, Guangzhou 510006, Peoples R China
[4] Anhui Univ, Sch Artificial Intelligence, Hefei 230601, Peoples R China
[5] South China Normal Univ, Sch Comp Sci, Guangzhou, Peoples R China
基金
中国国家自然科学基金;
关键词
Cross-modal hashing retrieval; Compatible triplet; Label network; Fusion attention;
D O I
10.1016/j.neucom.2024.128293
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Cross-modal hashing retrieval has emerged as a promising approach due to its advantages in storage efficiency and query speed for handling diverse multimodal data. However, existing cross-modal hashing retrieval methods often oversimplify similarity by solely considering identical labels across modalities and are sensitive to noise in the original multimodal data. To tackle this challenge, we propose a cross-modal hashing retrieval approach with compatible triplet representation. In the proposed approach, we integrate the essential feature representations and semantic information from text and images into their corresponding multi-label feature representations, and introduce a fusion attention module to extract text and image modalities with channel and spatial attention features, respectively, thereby enhancing compatible triplet-based semantic information in cross-modal hashing learning. Comprehensive experiments demonstrate the superiority of the proposed approach in retrieval accuracy compared to state-of-the-art methods on three public datasets.
引用
收藏
页数:9
相关论文
共 50 条
  • [1] Triplet Fusion Network Hashing for Unpaired Cross-Modal Retrieval
    Hu, Zhikai
    Liu, Xin
    Wang, Xingzhi
    Cheung, Yiu-ming
    Wang, Nannan
    Chen, Yewang
    ICMR'19: PROCEEDINGS OF THE 2019 ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, 2019, : 141 - 149
  • [2] Joint Coupled-Hashing Representation for Cross-Modal Retrieval
    Liu, Yihan
    Chen, Zhaojia
    Deng, Cheng
    Gao, Xinbo
    8TH INTERNATIONAL CONFERENCE ON INTERNET MULTIMEDIA COMPUTING AND SERVICE (ICIMCS2016), 2016, : 35 - 38
  • [3] Triplet-Based Deep Hashing Network for Cross-Modal Retrieval
    Deng, Cheng
    Chen, Zhaojia
    Liu, Xianglong
    Gao, Xinbo
    Tao, Dacheng
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2018, 27 (08) : 3893 - 3903
  • [4] Hashing for Cross-Modal Similarity Retrieval
    Liu, Yao
    Yuan, Yanhong
    Huang, Qiaoli
    Huang, Zhixing
    2015 11TH INTERNATIONAL CONFERENCE ON SEMANTICS, KNOWLEDGE AND GRIDS (SKG), 2015, : 1 - 8
  • [5] Online weighted hashing for cross-modal retrieval
    Jiang, Zining
    Weng, Zhenyu
    Li, Runhao
    Zhuang, Huiping
    Lin, Zhiping
    PATTERN RECOGNITION, 2025, 161
  • [6] Hierarchical Consensus Hashing for Cross-Modal Retrieval
    Sun, Yuan
    Ren, Zhenwen
    Hu, Peng
    Peng, Dezhong
    Wang, Xu
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 824 - 836
  • [7] Random Online Hashing for Cross-Modal Retrieval
    Jiang, Kaihang
    Wong, Wai Keung
    Fang, Xiaozhao
    Li, Jiaxing
    Qin, Jianyang
    Xie, Shengli
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2025, 36 (01) : 677 - 691
  • [8] Semantic consistency hashing for cross-modal retrieval
    Yao, Tao
    Kong, Xiangwei
    Fu, Haiyan
    Tian, Qi
    NEUROCOMPUTING, 2016, 193 : 250 - 259
  • [9] Fast Unmediated Hashing for Cross-Modal Retrieval
    Nie, Xiushan
    Liu, Xingbo
    Xi, Xiaoming
    Li, Chenglong
    Yin, Yilong
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2021, 31 (09) : 3669 - 3678
  • [10] Efficient Discriminative Hashing for Cross-Modal Retrieval
    Huang, Junfan
    Kang, Peipei
    Fang, Xiaozhao
    Han, Na
    Xie, Shengli
    Gao, Hongbo
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2024, 54 (06): : 3865 - 3878