Uncovering Distortion Differences: A Study of Adversarial Attacks and Machine Discriminability

被引:0
作者
Wang, Xiawei [1 ]
Li, Yao [2 ]
Hsieh, Cho-Jui [3 ]
Lee, Thomas C. M. [4 ]
机构
[1] Univ Calif Davis, Grad Grp Biostat, Davis, CA 95616 USA
[2] Univ North Carolina Chapel Hill, Dept Stat & Operat Res, Chapel Hill, NC 27599 USA
[3] Univ Calif Los Angeles, Dept Comp Sci, Los Angeles, CA 90095 USA
[4] Univ Calif Davis, Dept Stat, Davis, CA 95616 USA
基金
美国国家科学基金会;
关键词
Accuracy; Perturbation methods; Task analysis; Predictive models; Measurement; Reverse engineering; Visualization; Decision-based attacks; deep neural networks; gradient-based attacks; image classification; score-based attacks; ROBUSTNESS;
D O I
10.1109/ACCESS.2024.3446834
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep neural networks have performed remarkably in many areas, including image-related classification tasks. However, various studies have shown that they are vulnerable to adversarial examples - images carefully crafted to fool well-trained deep neural networks by introducing imperceptible perturbations to the original images. To better understand the inherent characteristics of adversarial attacks, this paper studies the features of three common attack families: gradient-based, score-based, and decision-based. The primary objective is to recognize distinct types of adversarial examples, as identifying the type of information possessed by the attacker can aid in developing effective defense strategies. This paper demonstrates that adversarial images from different attack families can be successfully identified with a simple model. To further investigate the reason behind the observations, this paper conducts carefully designed experiments to study the distortion patterns of different attacks. Experimental results on CIFAR10 and Tiny ImageNet validated the differences in distortion patterns between various attack types for both L-2 and L-infinity norm.
引用
收藏
页码:117872 / 117883
页数:12
相关论文
共 35 条
[1]   Square Attack: A Query-Efficient Black-Box Adversarial Attack via Random Search [J].
Andriushchenko, Maksym ;
Croce, Francesco ;
Flammarion, Nicolas ;
Hein, Matthias .
COMPUTER VISION - ECCV 2020, PT XXIII, 2020, 12368 :484-501
[2]  
Brendel W, 2018, Arxiv, DOI arXiv:1712.04248
[3]   Towards Evaluating the Robustness of Neural Networks [J].
Carlini, Nicholas ;
Wagner, David .
2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2017, :39-57
[4]   HopSkipJumpAttack: A Query-Efficient Decision-Based Attack [J].
Chen, Jianbo ;
Jordan, Michael, I ;
Wainwright, Martin J. .
2020 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP 2020), 2020, :1277-1294
[5]  
Chen PY, 2018, AAAI CONF ARTIF INTE, P10
[6]  
Chen PY, 2017, PROCEEDINGS OF THE 10TH ACM WORKSHOP ON ARTIFICIAL INTELLIGENCE AND SECURITY, AISEC 2017, P15, DOI 10.1145/3128572.3140448
[7]  
Cheng MH, 2020, Arxiv, DOI arXiv:1909.10773
[8]  
Croce F, 2019, 25TH AMERICAS CONFERENCE ON INFORMATION SYSTEMS (AMCIS 2019)
[9]  
Croce F, 2020, PR MACH LEARN RES, V119
[10]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848