Discrete asymmetric zero-shot hashing with application to cross-modal retrieval

被引:24
作者
Shu, Zhenqiu [1 ]
Yong, Kailing [1 ]
Yu, Jun [2 ]
Gao, Shengxiang [1 ]
Mao, Cunli [1 ]
Yu, Zhengtao [1 ]
机构
[1] Kunming Univ Sci & Technol, Sch Fac Informat Engn & Automation, Kunming, Peoples R China
[2] Zhengzhou Univ Light Ind, Coll Comp & Commun Engn, Zhengzhou, Peoples R China
关键词
Zero -shot hashing; Asymmetric; Cross -modal retrieval; Class attributes; Pairwise similarity;
D O I
10.1016/j.neucom.2022.09.037
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In recent years, cross-modal retrieval technology has attracted extensive attention with the massive growth of multimedia data. However, most cross-modal hashing methods mainly focus on exploring the retrieval of seen classes, while ignoring the retrieval of unseen classes. Therefore, traditional cross -modal hashing methods cannot achieve satisfactory performances in zero-shot retrieval. To mitigate this challenge, in this paper, we propose a novel zero-shot cross-modal retrieval method called discrete asym-metric zero-shot hashing(DAZSH), which fully exploits the supervised knowledge of multimodal data. Specifically, it integrates pairwise similarity, class attributes and semantic labels to guide zero-shot hash-ing learning. Moreover, our proposed DAZSH method combines the data features with the class attributes to obtain a semantic category representation for each category. Therefore, the relationships between seen and unseen classes can be effectively captured by learning a category representation vector for each instance. Therefore, the supervised knowledge can be transferred from the seen classes to the unseen classes. In addition, we develop an efficient discrete optimization strategy to solve the proposed model. Massive experiments on three benchmark datasets show that our proposed approach has achieved promising results in cross-modal retrieval tasks. The source code of this paper can be obtained from https://github.com/szq0816/DAZSH.(c) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页码:366 / 379
页数:14
相关论文
共 50 条
  • [41] Discrete online cross-modal hashing
    Zhan, Yu-Wei
    Wang, Yongxin
    Sun, Yu
    Wu, Xiao-Ming
    Luo, Xin
    Xu, Xin-Shun
    PATTERN RECOGNITION, 2022, 122
  • [42] Zero-shot Hashing with orthogonal projection for image retrieval
    Zhang, Haofeng
    Long, Yang
    Shao, Ling
    PATTERN RECOGNITION LETTERS, 2019, 117 : 201 - 209
  • [43] Online weighted hashing for cross-modal retrieval
    Jiang, Zining
    Weng, Zhenyu
    Li, Runhao
    Zhuang, Huiping
    Lin, Zhiping
    PATTERN RECOGNITION, 2025, 161
  • [44] Fast Unmediated Hashing for Cross-Modal Retrieval
    Nie, Xiushan
    Liu, Xingbo
    Xi, Xiaoming
    Li, Chenglong
    Yin, Yilong
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2021, 31 (09) : 3669 - 3678
  • [45] Semantic consistency hashing for cross-modal retrieval
    Yao, Tao
    Kong, Xiangwei
    Fu, Haiyan
    Tian, Qi
    NEUROCOMPUTING, 2016, 193 : 250 - 259
  • [46] Hierarchical Consensus Hashing for Cross-Modal Retrieval
    Sun, Yuan
    Ren, Zhenwen
    Hu, Peng
    Peng, Dezhong
    Wang, Xu
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 824 - 836
  • [47] Random Online Hashing for Cross-Modal Retrieval
    Jiang, Kaihang
    Wong, Wai Keung
    Fang, Xiaozhao
    Li, Jiaxing
    Qin, Jianyang
    Xie, Shengli
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2025, 36 (01) : 677 - 691
  • [48] Efficient Discriminative Hashing for Cross-Modal Retrieval
    Huang, Junfan
    Kang, Peipei
    Fang, Xiaozhao
    Han, Na
    Xie, Shengli
    Gao, Hongbo
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2024, 54 (06): : 3865 - 3878
  • [49] GrowBit: Incremental Hashing for Cross-Modal Retrieval
    Mandal, Devraj
    Annadani, Yashas
    Biswas, Soma
    COMPUTER VISION - ACCV 2018, PT IV, 2019, 11364 : 305 - 321
  • [50] Two-Stage Asymmetric Similarity Preserving Hashing for Cross-Modal Retrieval
    Huang, Junfan
    Kang, Peipei
    Han, Na
    Chen, Yonghao
    Fang, Xiaozhao
    Gao, Hongbo
    Zhou, Guoxu
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (01) : 429 - 444