Boosting the Adversarial Transferability of Surrogate Models with Dark Knowledge

被引:0
|
作者
Yang, Dingcheng [1 ,2 ]
Xiao, Zihao [2 ]
Yu, Wenjian [1 ]
机构
[1] Tsinghua Univ, Dept Comp Sci Tech, BNRist, Beijing, Peoples R China
[2] RealAI, Beijing, Peoples R China
关键词
Deep learning; Image classification; Black-box adversarial attack; Transfer-based attack; Dark knowledge;
D O I
10.1109/ICTAI59109.2023.00098
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks (DNNs) are vulnerable to adversarial examples. And, the adversarial examples have transferability, which means that an adversarial example for a DNN model can fool another model with a non-trivial probability. This gave birth to the transfer-based attack where the adversarial examples generated by a ate model are used to conduct black-box attacks. There are some work on generating the adversarial examples from a given surrogate model with better transferability. However, training a special surrogate model to generate adversarial examples with better transferability is relatively under-explored. This paper proposes a method for training a surrogate model with dark knowledge to boost the transferability of the adversarial examples generated by the surrogate model. This trained surrogate model is named dark surrogate model (DSM). The proposed method for training a DSM consists of two key components: a teacher model extracting dark knowledge, and the mixing augmentation skill enhancing dark knowledge of training data. We conducted extensive experiments to show that the proposed method can substantially improve the adversarial transferability of surrogate models across different architectures of surrogate models and optimizers for generating adversarial examples, and it can be applied to other scenarios of transfer-based attack that contain dark knowledge, like face verification. Our code is publicly available at https://github.com/ydc123/Dark Surrogate Model.
引用
收藏
页码:627 / 635
页数:9
相关论文
共 50 条
  • [21] Boosting the Transferability of Video Adversarial Examples via Temporal Translation
    Wei, Zhipeng
    Chen, Jingjing
    Wu, Zuxuan
    Jiang, Yu-Gang
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 2659 - 2667
  • [22] Toward Understanding and Boosting Adversarial Transferability From a Distribution Perspective
    Zhu, Yao
    Chen, Yuefeng
    Li, Xiaodan
    Chen, Kejiang
    He, Yuan
    Tian, Xiang
    Zheng, Bolun
    Chen, Yaowu
    Huang, Qingming
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2022, 31 : 6487 - 6501
  • [23] The Ultimate Combo: Boosting Adversarial Example Transferability by Composing Data Augmentations
    Yun, Zebin
    Weingarten, Achi-Or
    Ronen, Eyal
    Sharif, Mahmood
    PROCEEDINGS OF THE 2024 WORKSHOP ON ARTIFICIAL INTELLIGENCE AND SECURITY, AISEC 2024, 2024, : 113 - 124
  • [24] Boosting Adversarial Transferability via Logits Mixup With Dominant Decomposed Feature
    Weng, Juanjuan
    Luo, Zhiming
    Li, Shaozi
    Lin, Dazhen
    Zhong, Zhun
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 8939 - 8951
  • [25] Boosting Transferability of Targeted Adversarial Examples via Hierarchical Generative Networks
    Yang, Xiao
    Dong, Yinpeng
    Pang, Tianyu
    Su, Hang
    Zhu, Jun
    COMPUTER VISION - ECCV 2022, PT IV, 2022, 13664 : 725 - 742
  • [26] Boosting Adversarial Transferability with Shallow-Feature Attack on SAR Images
    Lin, Gengyou
    Pan, Zhisong
    Zhou, Xingyu
    Duan, Yexin
    Bai, Wei
    Zhan, Dazhi
    Zhu, Leqian
    Zhao, Gaoqiang
    Li, Tao
    REMOTE SENSING, 2023, 15 (10)
  • [27] Boosting the transferability of adversarial attacks with adaptive points selecting in temporal neighborhood
    Zhu, Hegui
    Zheng, Haoran
    Zhu, Ying
    Sui, Xiaoyan
    INFORMATION SCIENCES, 2023, 641
  • [28] LGV: Boosting Adversarial Example Transferability from Large Geometric Vicinity
    Gubri, Martin
    Cordy, Maxime
    Papadakis, Mike
    Le Traon, Yves
    Sen, Koushik
    COMPUTER VISION - ECCV 2022, PT IV, 2022, 13664 : 603 - 618
  • [29] Set-level Guidance Attack: Boosting Adversarial Transferability of Vision-Language Pre-training Models
    Lu, Dong
    Wang, Zhiqiang
    Wang, Teng
    Guan, Weili
    Gao, Hongchang
    Zheng, Feng
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, : 102 - 111
  • [30] Probability-Distribution-Guided Adversarial Sample Attacks for Boosting Transferability and Interpretability
    Li, Hongying
    Yu, Miaomiao
    Li, Xiaofei
    Zhang, Jun
    Li, Shuohao
    Lei, Jun
    Huang, Hairong
    MATHEMATICS, 2023, 11 (13)