WFormer: A Transformer-Based Soft Fusion Model for Robust Image Watermarking

被引:3
|
作者
Luo, Ting [1 ]
Wu, Jun [2 ]
He, Zhouyan [1 ]
Xu, Haiyong [2 ]
Jiang, Gangyi [2 ]
Chang, Chin-Chen [3 ]
机构
[1] Ningbo Univ, Coll Sci & Technol, Ningbo 315212, Peoples R China
[2] Ningbo Univ, Fac Informat Sci & Engn, Ningbo 315211, Peoples R China
[3] Feng Chia Univ, Dept Informat Engn & Comp Sci, Taichung 40724, Taiwan
基金
中国国家自然科学基金;
关键词
Watermarking; Feature extraction; Transformers; Decoding; Convolution; Noise; Robustness; transformer; soft fusion; cross-attention;
D O I
10.1109/TETCI.2024.3386916
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Most deep neural network (DNN) based image watermarking models often employ the encoder-noise-decoder structure, in which watermark is simply duplicated for expansion and then directly fused with image features to produce the encoded image. However, simple duplication will generate watermark over-redundancies, and the communication between the cover image and watermark in different domains is lacking in image feature extraction and direction fusion, which degrades the watermarking performance. To solve those drawbacks, this paper proposes a Transformer-based soft fusion model for robust image watermarking, namely WFormer. Specifically, to expand watermark effectively, a watermark preprocess module (WPM) is designed with Transformers to extract valid and expanded watermark features by computing its self-attention. Then, to replace direct fusion, a soft fusion module (SFM) is deployed to integrate Transformers into image fusion with watermark by mining their long-range correlations. Precisely, self-attention is computed to extract their own latent features, and meanwhile, cross-attention is learned for bridging their gap to embed watermark effectively. In addition, a feature enhancement module (FEM) builds communication between the cover image and watermark by capturing their cross-feature dependencies, which tunes image features in accordance with watermark features for better fusion. Experimental results show that the proposed WFormer outperforms the existing state-of-the-art watermarking models in terms of invisibility, robustness, and embedding capacity. Furthermore, ablation results prove the effectiveness of the WPM, the FEM, and the SFM.
引用
收藏
页码:1 / 18
页数:18
相关论文
共 50 条
  • [21] TMCIH: Perceptual Robust Image Hashing with Transformer-based Multi-layer Constraints
    Fang, Yaodong
    Zhou, Yuanding
    Li, Xinran
    Kong, Ping
    Qin, Chuan
    PROCEEDINGS OF THE 2023 ACM WORKSHOP ON INFORMATION HIDING AND MULTIMEDIA SECURITY, IH&MMSEC 2023, 2023, : 7 - 12
  • [22] Image-Text Person Re-Identification with Transformer-Based Modal Fusion
    Li, Xin
    Guo, Hubo
    Zhang, Meiling
    Fu, Bo
    ELECTRONICS, 2025, 14 (03):
  • [23] Aware-Transformer: A Novel Pure Transformer-Based Model for Remote Sensing Image Captioning
    Cao, Yukun
    Yan, Jialuo
    Tang, Yijia
    He, Zhenyi
    Xu, Kangle
    Cheng, Yu
    ADVANCES IN COMPUTER GRAPHICS, CGI 2023, PT I, 2024, 14495 : 105 - 117
  • [24] Robust Transformer-based model for spatiotemporal PM2.5 prediction in California
    Tong, Weitian
    Limperis, Jordan
    Hamza-Lup, Felix
    Xu, Yao
    Li, Lixin
    EARTH SCIENCE INFORMATICS, 2024, 17 (01) : 315 - 328
  • [25] A robust digital image watermarking method using wavelet-based fusion
    Kundur, D
    Hatzinakos, D
    INTERNATIONAL CONFERENCE ON IMAGE PROCESSING - PROCEEDINGS, VOL I, 1997, : 544 - 547
  • [26] A robust color image watermarking scheme in the fusion domain based on LU factorization
    Su, Qingtang
    Sun, Yehan
    Xia, Yu
    Wang, Zengfeng
    OPTICS AND LASER TECHNOLOGY, 2024, 174
  • [27] TransMF: Transformer-Based Multi-Scale Fusion Model for Crack Detection
    Ju, Xiaochen
    Zhao, Xinxin
    Qian, Shengsheng
    MATHEMATICS, 2022, 10 (13)
  • [28] A Sparse Transformer-Based Approach for Image Captioning
    Lei, Zhou
    Zhou, Congcong
    Chen, Shengbo
    Huang, Yiyong
    Liu, Xianrui
    IEEE Access, 2020, 8 : 213437 - 213446
  • [29] A Sparse Transformer-Based Approach for Image Captioning
    Lei, Zhou
    Zhou, Congcong
    Chen, Shengbo
    Huang, Yiyong
    Liu, Xianrui
    IEEE ACCESS, 2020, 8 : 213437 - 213446
  • [30] Transformer-based Extraction of Deep Image Models
    Battis, Verena
    Penner, Alexander
    2022 IEEE 7TH EUROPEAN SYMPOSIUM ON SECURITY AND PRIVACY (EUROS&P 2022), 2022, : 320 - 336