WFormer: A Transformer-Based Soft Fusion Model for Robust Image Watermarking

被引:3
|
作者
Luo, Ting [1 ]
Wu, Jun [2 ]
He, Zhouyan [1 ]
Xu, Haiyong [2 ]
Jiang, Gangyi [2 ]
Chang, Chin-Chen [3 ]
机构
[1] Ningbo Univ, Coll Sci & Technol, Ningbo 315212, Peoples R China
[2] Ningbo Univ, Fac Informat Sci & Engn, Ningbo 315211, Peoples R China
[3] Feng Chia Univ, Dept Informat Engn & Comp Sci, Taichung 40724, Taiwan
基金
中国国家自然科学基金;
关键词
Watermarking; Feature extraction; Transformers; Decoding; Convolution; Noise; Robustness; transformer; soft fusion; cross-attention;
D O I
10.1109/TETCI.2024.3386916
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Most deep neural network (DNN) based image watermarking models often employ the encoder-noise-decoder structure, in which watermark is simply duplicated for expansion and then directly fused with image features to produce the encoded image. However, simple duplication will generate watermark over-redundancies, and the communication between the cover image and watermark in different domains is lacking in image feature extraction and direction fusion, which degrades the watermarking performance. To solve those drawbacks, this paper proposes a Transformer-based soft fusion model for robust image watermarking, namely WFormer. Specifically, to expand watermark effectively, a watermark preprocess module (WPM) is designed with Transformers to extract valid and expanded watermark features by computing its self-attention. Then, to replace direct fusion, a soft fusion module (SFM) is deployed to integrate Transformers into image fusion with watermark by mining their long-range correlations. Precisely, self-attention is computed to extract their own latent features, and meanwhile, cross-attention is learned for bridging their gap to embed watermark effectively. In addition, a feature enhancement module (FEM) builds communication between the cover image and watermark by capturing their cross-feature dependencies, which tunes image features in accordance with watermark features for better fusion. Experimental results show that the proposed WFormer outperforms the existing state-of-the-art watermarking models in terms of invisibility, robustness, and embedding capacity. Furthermore, ablation results prove the effectiveness of the WPM, the FEM, and the SFM.
引用
收藏
页码:1 / 18
页数:18
相关论文
共 50 条
  • [41] Transformer-based fusion model for mild depression recognition with EEG and pupil area signals
    Zhu, Jing
    Li, Yuanlong
    Yang, Changlin
    Cai, Hanshu
    Li, Xiaowei
    Hu, Bin
    MEDICAL & BIOLOGICAL ENGINEERING & COMPUTING, 2025,
  • [42] Identification of lesion bioactivity in hepatic cystic echinococcosis using a transformer-based fusion model
    Wang, Zhanjin
    Li, Fuyuan
    Cai, Junjie
    Xue, Zhangtuo
    Du, Kaihao
    Tao, Yongping
    Zhang, Hanxi
    Zhou, Ying
    Fan, Haining
    Wang, Zhan
    JOURNAL OF INFECTION, 2025, 90 (04)
  • [43] CityTFT: A temporal fusion transformer-based surrogate model for urban building energy modeling
    Dai, Ting-Yu
    Niyogi, Dev
    Nagy, Zoltan
    APPLIED ENERGY, 2025, 389
  • [44] A Robust Image Watermarking Technique Based On Image Interlacing
    Ibrahim, Mohamed M.
    Kader, Neamat S. Abdel
    Zorkany, M.
    2014 31ST NATIONAL RADIO SCIENCE CONFERENCE (NRSC), 2014, : 92 - 98
  • [45] DesnowFormer: an effective transformer-based image desnowing network
    Zhang, Ting
    Jiang, Nanfeng
    Lin, Junhong
    Lin, Jielian
    Zhao, Tiesong
    2022 IEEE INTERNATIONAL CONFERENCE ON VISUAL COMMUNICATIONS AND IMAGE PROCESSING (VCIP), 2022,
  • [46] Recent progress in transformer-based medical image analysis
    Liu, Zhaoshan
    Lv, Qiujie
    Yang, Ziduo
    Li, Yifan
    Lee, Chau Hung
    Shen, Lei
    COMPUTERS IN BIOLOGY AND MEDICINE, 2023, 164
  • [47] TransInpaint: Transformer-based Image Inpainting with Context Adaptation
    Shamsolmoali, Pourya
    Zareapoor, Masoumeh
    Granger, Eric
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, ICCVW, 2023, : 849 - 858
  • [48] A Novel Transformer-Based Attention Network for Image Dehazing
    Gao, Guanlei
    Cao, Jie
    Bao, Chun
    Hao, Qun
    Ma, Aoqi
    Li, Gang
    SENSORS, 2022, 22 (09)
  • [49] Transformer-based image generation from scene graphs
    Sortino, Renato
    Palazzo, Simone
    Rundo, Francesco
    Spampinato, Concetto
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2023, 233
  • [50] Symmetric transformer-based network for unsupervised image registration
    Ma, Mingrui
    Xu, Yuanbo
    Song, Lei
    Liu, Guixia
    KNOWLEDGE-BASED SYSTEMS, 2022, 257