A hypothetical defenses-based training framework for generating transferable adversarial examples

被引:0
|
作者
Hao, Lingguang [1 ]
Hao, Kuangrong [1 ]
Jin, Yaochu [2 ]
Zhao, Hongzhi [1 ]
机构
[1] Donghua Univ, Coll Informat Sci & Technol, Shanghai 201620, Peoples R China
[2] Westlake Univ, Sch Engn, Trustworthy & Gen AI Lab, Hangzhou 310024, Peoples R China
基金
中国国家自然科学基金;
关键词
Transferable adversarial attack; Hypothetical defense; Image denoising; Adversarial learning;
D O I
10.1016/j.knosys.2024.112602
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Transfer-based attacks utilize the proxy model to craft adversarial examples against the target model and make significant advancements in the realm of black-box attacks. Recent research suggests that these attacks can be enhanced by incorporating adversarial defenses into the training process of adversarial examples. Specifically, adversarial defenses supervise the training process, forcing the attacker to overcome greater challenges and produce more robust adversarial examples with enhanced transferability. However, current methods mainly rely on limited input transformation defenses, which apply only linear affine changes. These defenses are insufficient for effectively removing harmful content from adversarial examples, resulting in restricted improvements in their transferability. To address this issue, we propose a novel training framework named Transfer-based Attacks through Hypothesis Defense (TA-HD). This framework enhances the generalization of adversarial examples by integrating a hypothesis defense mechanism into the proxy model. Specifically, we propose an input denoising network as the hypothesis defense to effectively remove harmful noise from adversarial examples. Furthermore, we introduce an adversarial training strategy and design specific adversarial loss functions to optimize the input denoising network's parameters. The visualization of the training process demonstrates the effective denoising capability of the hypothesized defense mechanism and the stability of the training process. Extensive experiments show that the proposed training framework significantly improves the success rate of transfer-based attacks by up to 19.9%. The code is available at https://github.com/haolingguang/TA-HD.
引用
收藏
页数:13
相关论文
共 50 条
  • [1] Efficient Adversarial Training with Transferable Adversarial Examples
    Zheng, Haizhong
    Zhang, Ziqi
    Gu, Juncheng
    Lee, Honglak
    Prakash, Atul
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 1178 - 1187
  • [2] Generating Transferable Adversarial Examples for Speech Classification
    Kim, Hoki
    Park, Jinseong
    Lee, Jaewook
    PATTERN RECOGNITION, 2023, 137
  • [3] Generating transferable adversarial examples based on perceptually-aligned perturbation
    Chen, Hongqiao
    Lu, Keda
    Wang, Xianmin
    Li, Jin
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2021, 12 (11) : 3295 - 3307
  • [4] Generating transferable adversarial examples based on perceptually-aligned perturbation
    Hongqiao Chen
    Keda Lu
    Xianmin Wang
    Jin Li
    International Journal of Machine Learning and Cybernetics, 2021, 12 : 3295 - 3307
  • [5] Common knowledge learning for generating transferable adversarial examples
    Yang, Ruijie
    Guo, Yuanfang
    Wang, Junfu
    Zhou, Jiantao
    Wang, Yunhong
    FRONTIERS OF COMPUTER SCIENCE, 2025, 19 (10)
  • [6] Generating Transferable Adversarial Examples against Vision Transformers
    Wang, Yuxuan
    Wang, Jiakai
    Yin, Zinxin
    Gong, Ruihao
    Wang, Jingyi
    Liu, Aishan
    Liu, Xianglong
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 5181 - 5190
  • [7] Evading Defenses to Transferable Adversarial Examples by Translation-Invariant Attacks
    Dong, Yinpeng
    Pang, Tianyu
    Su, Hang
    Zhu, Jun
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 4307 - 4316
  • [8] Generating Transferable Adversarial Examples From the Perspective of Ensemble and Distribution
    Zhang, Huangyi
    Liu, Ximeng
    PROCEEDINGS OF 2024 3RD INTERNATIONAL CONFERENCE ON CYBER SECURITY, ARTIFICIAL INTELLIGENCE AND DIGITAL ECONOMY, CSAIDE 2024, 2024, : 173 - 177
  • [9] Feature-Based Adversarial Training for Deep Learning Models Resistant to Transferable Adversarial Examples
    Ryu, Gwonsang
    Choi, Daeseon
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2022, E105D (05) : 1039 - 1049
  • [10] An efficient framework for generating robust adversarial examples
    Zhang, Lili
    Wang, Xiaoping
    Lu, Kai
    Peng, Shaoliang
    Wang, Xiaodong
    INTERNATIONAL JOURNAL OF INTELLIGENT SYSTEMS, 2020, 35 (09) : 1433 - 1449