Your Attack Is Too DUMB: Formalizing Attacker Scenarios for Adversarial Transferability

被引:4
作者
Alecci, Marco [1 ]
Conti, Mauro [2 ]
Marchiori, Francesco [2 ]
Martinelli, Luca [2 ]
Pajola, Luca [2 ]
机构
[1] Univ Luxembourg, SnT, Luxembourg, Luxembourg
[2] Univ Padua, Padua, Italy
来源
PROCEEDINGS OF THE 26TH INTERNATIONAL SYMPOSIUM ON RESEARCH IN ATTACKS, INTRUSIONS AND DEFENSES, RAID 2023 | 2023年
关键词
Adversarial Machine Learning; Adversarial Attacks; Evasion Attacks; Transferability; Surrogate Model;
D O I
10.1145/3607199.3607227
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Evasion attacks are a threat to machine learning models, where adversaries attempt to affect classifiers by injecting malicious samples. An alarming side-effect of evasion attacks is their ability to transfer among different models: this property is called transferability. Therefore, an attacker can produce adversarial samples on a custom model (surrogate) to conduct the attack on a victim's organization later. Although literature widely discusses how adversaries can transfer their attacks, their experimental settings are limited and far from reality. For instance, many experiments consider both attacker and defender sharing the same dataset, balance level (i.e., how the ground truth is distributed), and model architecture. In this work, we propose the DUMB attacker model. This framework allows analyzing if evasion attacks fail to transfer when the training conditions of surrogate and victim models differ. DUMB considers the following conditions: Dataset soUrces, Model architecture, and the Balance of the ground truth. We then propose a novel testbed to evaluate many state-of-the-art evasion attacks with DUMB; the testbed consists of three computer vision tasks with two distinct datasets each, four types of balance levels, and three model architectures. Our analysis, which generated 13K tests over 14 distinct attacks, led to numerous novel findings in the scope of transferable attacks with surrogate models. In particular, mismatches between attackers and victims in terms of dataset source, balance levels, and model architecture lead to non-negligible loss of attack performance.
引用
收藏
页码:315 / 329
页数:15
相关论文
共 33 条
  • [1] Andriushchenko Maksym, 2020, Computer Vision - ECCV 2020. 16th European Conference. Proceedings. Lecture Notes in Computer Science (LNCS 12368), P484, DOI 10.1007/978-3-030-58592-1_29
  • [2] Apruzzese G, 2022, Arxiv, DOI arXiv:2212.14315
  • [3] Apruzzese G, 2018, 2018 IEEE 17TH INTERNATIONAL SYMPOSIUM ON NETWORK COMPUTING AND APPLICATIONS (NCA)
  • [4] Barreno M., 2006, P ACM S INF COMP COM, P16, DOI DOI 10.1145/1128817.1128824
  • [5] SMOTE: Synthetic minority over-sampling technique
    Chawla, Nitesh V.
    Bowyer, Kevin W.
    Hall, Lawrence O.
    Kegelmeyer, W. Philip
    [J]. 2002, American Association for Artificial Intelligence (16)
  • [6] Conti Mauro, 2022, arXiv
  • [7] Davidson T, 2017, P INT AAAI C WEB SOC, P1, DOI [DOI 10.1609/ICWSM.V11I1.14955, 10.1609/icwsm.v11i1.14955]
  • [8] Demontis A, 2019, PROCEEDINGS OF THE 28TH USENIX SECURITY SYMPOSIUM, P321
  • [9] Evading Defenses to Transferable Adversarial Examples by Translation-Invariant Attacks
    Dong, Yinpeng
    Pang, Tianyu
    Su, Hang
    Zhu, Jun
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 4307 - 4316
  • [10] Frid-Adar M, 2018, I S BIOMED IMAGING, P289, DOI 10.1109/ISBI.2018.8363576