Boosting the Adversarial Transferability of Surrogate Models with Dark Knowledge

被引:0
|
作者
Yang, Dingcheng [1 ,2 ]
Xiao, Zihao [2 ]
Yu, Wenjian [1 ]
机构
[1] Tsinghua Univ, Dept Comp Sci Tech, BNRist, Beijing, Peoples R China
[2] RealAI, Beijing, Peoples R China
关键词
Deep learning; Image classification; Black-box adversarial attack; Transfer-based attack; Dark knowledge;
D O I
10.1109/ICTAI59109.2023.00098
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks (DNNs) are vulnerable to adversarial examples. And, the adversarial examples have transferability, which means that an adversarial example for a DNN model can fool another model with a non-trivial probability. This gave birth to the transfer-based attack where the adversarial examples generated by a ate model are used to conduct black-box attacks. There are some work on generating the adversarial examples from a given surrogate model with better transferability. However, training a special surrogate model to generate adversarial examples with better transferability is relatively under-explored. This paper proposes a method for training a surrogate model with dark knowledge to boost the transferability of the adversarial examples generated by the surrogate model. This trained surrogate model is named dark surrogate model (DSM). The proposed method for training a DSM consists of two key components: a teacher model extracting dark knowledge, and the mixing augmentation skill enhancing dark knowledge of training data. We conducted extensive experiments to show that the proposed method can substantially improve the adversarial transferability of surrogate models across different architectures of surrogate models and optimizers for generating adversarial examples, and it can be applied to other scenarios of transfer-based attack that contain dark knowledge, like face verification. Our code is publicly available at https://github.com/ydc123/Dark Surrogate Model.
引用
收藏
页码:627 / 635
页数:9
相关论文
共 50 条
  • [41] Boosting the Transferability of Adversarial Examples with Gradient-Aligned Ensemble Attack for Speaker Recognition
    Li, Zhuhai
    Zhang, Jie
    Guo, Wu
    Wu, Haochen
    INTERSPEECH 2024, 2024, : 532 - 536
  • [42] Enhancing Transferability of Adversarial Examples by Successively Attacking Multiple Models
    Zhang, Xiaolin
    Zhang, Wenwen
    Liu, Lixin
    Wang, Yongping
    Gao, Lu
    Zhang, Shuai
    International Journal of Network Security, 2023, 25 (02) : 306 - 316
  • [43] Improving Transferability of Adversarial Patches on Face Recognition with Generative Models
    Xiao, Zihao
    Gao, Xianfeng
    Fu, Chilin
    Dong, Yinpeng
    Gao, Wei
    Zhang, Xiaolu
    Zhou, Jun
    Zhu, Jun
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 11840 - 11849
  • [44] Surrogate models for direct dark matter detection
    Cerdeno, D. G.
    Cheek, A.
    Reid, E.
    Schulz, H.
    JOURNAL OF COSMOLOGY AND ASTROPARTICLE PHYSICS, 2018, (08):
  • [45] FedGTST: Boosting Global Transferability of Federated Models via Statistics Tuning
    Ma, Evelyn
    Etesami, Rasoul
    Pan, Chao
    Zhao, Han
    Milenkovic, Olgica
    arXiv,
  • [46] Improving the Transferability of Adversarial Samples with Adversarial Transformations
    Wu, Weibin
    Su, Yuxin
    Lyu, Michael R.
    King, Irwin
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 9020 - 9029
  • [47] Ranking the Transferability of Adversarial Examples
    Levy, Moshe
    Amit, Guy
    Elovici, Yuval
    Mirsky, Yisroel
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2024, 15 (05)
  • [48] Exploring Transferability on Adversarial Attacks
    Alvarez, Enrique
    Alvarez, Rafael
    Cazorla, Miguel
    IEEE ACCESS, 2023, 11 : 105545 - 105556
  • [49] Boosting Transferability in Vision-Language Attacks via Diversification Along the Intersection Region of Adversarial Trajectory
    Gao, Sensen
    Jia, Xiaojun
    Rene, Xuhong
    Tsang, Ivor
    Guo, Qing
    COMPUTER VISION-ECCV 2024, PT LVII, 2025, 15115 : 442 - 460
  • [50] Closer Look at the Transferability of Adversarial Examples: How They Fool Different Models Differently
    Waseda, Futa
    Nishikawa, Sosuke
    Trung-Nghia Le
    Nguyen, Huy H.
    Echizen, Isao
    2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2023, : 1360 - 1368