Unsupervised Contrastive Hashing With Autoencoder Semantic Similarity for Cross-Modal Retrieval in Remote Sensing

被引:0
|
作者
Liu, Na [1 ]
Wu, Guodong [2 ]
Huang, Yonggui [3 ]
Chen, Xi [4 ]
Li, Qingdu [1 ]
Wan, Lihong [2 ]
机构
[1] Univ Shanghai Sci & Technol, Shanghai 200093, Peoples R China
[2] Origin Dynam Intelligent Robot Co Ltd, Zhengzhou 450000, Peoples R China
[3] Peking Univ, Big Data Res Ctr, Beijing 100091, Peoples R China
[4] East China Normal Univ, Software Engn Inst, Shanghai 200062, Peoples R China
关键词
Cross-modal retrieval; hashing; Remote sensing; unsupervised contrastive learning; remote sensing; IMAGE RETRIEVAL; NETWORK;
D O I
10.1109/JSTARS.2025.3538701
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In large-scale multimodal remote sensing data archives, the application of cross-modal technology to achieve fast retrieval between different modalities has attracted great attention. In this article, we focus on cross-modal retrieval technology between remote sensing images and text. At present, there is still a large heterogeneity problem in the semantic information extracted from different modal data in the remote sensing field, which leads to the inability to effectively utilize intraclass similarities and interclass differences in the hash learning process, ultimately resulting in low cross-modal retrieval accuracy. In addition, supervised learning-based methods require a large number of labeled training samples, which limits the large-scale application of hash-based cross-modal retrieval technology in the remote sensing field. To address this problem, this article proposes a new unsupervised cross-autoencoder contrast hashing method for RS retrieval. This method constructs an end-to-end deep hashing model, which mainly includes a feature extraction module and a hash representation module. The feature extraction module is mainly responsible for extracting deep semantic information from different modal data and sends the different modal semantic information to the hash representation module through the intermediate layer to learn and generate binary hash codes. In the hashing module, we introduce a new multiobjective loss function to increase the expression of intramodal and intermodal semantic consistency through multiscale semantic similarity constraints and contrastive learning and add a cross-autoencoding module to reconstruct and compare hash features to reduce the loss of semantic information during the learning process. This article conducts a large number of experiments on the UC Merced Land dataset and the RSICD dataset. The experimental results of these two popular benchmark datasets show that the proposed CACH method outperforms the most advanced unsupervised cross-modal hashing methods in RS.
引用
收藏
页码:6047 / 6059
页数:13
相关论文
共 50 条
  • [41] Deep Adaptively-Enhanced Hashing With Discriminative Similarity Guidance for Unsupervised Cross-Modal Retrieval
    Shi, Yufeng
    Zhao, Yue
    Liu, Xin
    Zheng, Feng
    Ou, Weihua
    You, Xinge
    Peng, Qinmu
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (10) : 7255 - 7268
  • [42] Unsupervised deep hashing with multiple similarity preservation for cross-modal image-text retrieval
    Xiong, Siyu
    Pan, Lili
    Ma, Xueqiang
    Hu, Qinghua
    Beckman, Eric
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024, 15 (10) : 4423 - 4434
  • [43] Semantic-rebased cross-modal hashing for scalable unsupervised text-visual retrieval
    Wang, Weiwei
    Shen, Yuming
    Zhang, Haofeng
    Liu, Li
    INFORMATION PROCESSING & MANAGEMENT, 2020, 57 (06)
  • [44] Unsupervised Dual Hashing Coding (UDC) on Semantic Tagging and Sample Content for Cross-Modal Retrieval
    Cai, Hongmin
    Zhang, Bin
    Li, Junyu
    Hu, Bin
    Chen, Jiazhou
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 9109 - 9120
  • [45] Latent Semantic Sparse Hashing for Cross-Modal Similarity Search
    Zhou, Jile
    Ding, Guiguang
    Guo, Yuchen
    SIGIR'14: PROCEEDINGS OF THE 37TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2014, : 415 - 424
  • [46] Deep supervised fused similarity hashing for cross-modal retrieval
    Ng W.W.Y.
    Xu Y.
    Tian X.
    Wang H.
    Multimedia Tools and Applications, 2024, 83 (39) : 86537 - 86555
  • [47] Contrastive Incomplete Cross-Modal Hashing
    Luo, Haoyang
    Zhang, Zheng
    Nie, Liqiang
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (11) : 5823 - 5834
  • [48] Deep semantic hashing with dual attention for cross-modal retrieval
    Jiagao Wu
    Weiwei Weng
    Junxia Fu
    Linfeng Liu
    Bin Hu
    Neural Computing and Applications, 2022, 34 : 5397 - 5416
  • [49] Joint Semantic Preserving Sparse Hashing for Cross-Modal Retrieval
    Hu, Zhikai
    Cheung, Yiu-Ming
    Li, Mengke
    Lan, Weichao
    Zhang, Donglin
    Liu, Qiang
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (04) : 2989 - 3002
  • [50] Joint-Modal Graph Convolutional Hashing for unsupervised cross-modal retrieval
    Meng, Hui
    Zhang, Huaxiang
    Liu, Li
    Liu, Dongmei
    Lu, Xu
    Guo, Xinru
    NEUROCOMPUTING, 2024, 595