Is Robustness the Cost of Accuracy? - A Comprehensive Study on the Robustness of 18 Deep Image Classification Models

被引:183
|
作者
Su, Dong [1 ]
Zhang, Huan [2 ]
Chen, Hongge [3 ]
Yi, Jinfeng [4 ]
Chen, Pin-Yu [1 ]
Gao, Yupeng [1 ]
机构
[1] IBM Res, New York, NY 10598 USA
[2] Univ Calif Davis, Davis, CA 95616 USA
[3] MIT, Cambridge, MA 02139 USA
[4] JD AI Res, Beijing, Peoples R China
来源
COMPUTER VISION - ECCV 2018, PT XII | 2018年 / 11216卷
关键词
Deep neural networks; Adversarial attacks; Robustness;
D O I
10.1007/978-3-030-01258-8_39
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The prediction accuracy has been the long-lasting and sole standard for comparing the performance of different image classification models, including the ImageNet competition. However, recent studies have highlighted the lack of robustness in well-trained deep neural networks to adversarial examples. Visually imperceptible perturbations to natural images can easily be crafted and mislead the image classifiers towards misclassification. To demystify the trade-offs between robustness and accuracy, in this paper we thoroughly benchmark 18 ImageNet models using multiple robustness metrics, including the distortion, success rate and transferability of adversarial examples between 306 pairs of models. Our extensive experimental results reveal several new insights: (1) linear scaling law - the empirical l(2) and l(infinity) distortion metrics scale linearly with the logarithm of classification error; (2) model architecture is a more critical factor to robustness than model size, and the disclosed accuracy-robustness Pareto frontier can be used as an evaluation criterion for ImageNet model designers; (3) for a similar network architecture, increasing network depth slightly improves robustness in l(infinity) distortion; (4) there exist models (in VGG family) that exhibit high adversarial transferability, while most adversarial examples crafted from one model can only be transferred within the same family. Experiment code is publicly available at https://github.com/huanzhang12/Adversarial_Survey.
引用
收藏
页码:644 / 661
页数:18
相关论文
共 50 条
  • [31] The Impact of Model Variations on the Robustness of Deep Learning Models in Adversarial Settings
    Juraev, Firuz
    Abuhamad, Mohammed
    Woo, Simon S.
    Thiruvathukal, George K.
    Abuhmed, Tamer
    2024 SILICON VALLEY CYBERSECURITY CONFERENCE, SVCC 2024, 2024,
  • [32] ADVRET: An Adversarial Robustness Evaluating and Testing Platform for Deep Learning Models
    Ren, Fei
    Yang, Yonghui
    Hu, Chi
    Zhou, Yuyao
    Ma, Siyou
    2021 21ST INTERNATIONAL CONFERENCE ON SOFTWARE QUALITY, RELIABILITY AND SECURITY COMPANION (QRS-C 2021), 2021, : 9 - 14
  • [33] Adversarial Robustness for Deep Learning-Based Wildfire Prediction Models
    Ide, Ryo
    Yang, Lei
    FIRE-SWITZERLAND, 2025, 8 (02):
  • [34] Towards robustness evaluation of backdoor defense on quantized deep learning models
    Zhu, Yifan
    Peng, Huaibing
    Fu, Anmin
    Yang, Wei
    Ma, Hua
    Al-Sarawi, Said F.
    Abbott, Derek
    Gao, Yansong
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 255
  • [35] CMMR: A Composite Multidimensional Models Robustness Evaluation Framework for Deep Learning
    Liu, Wanyi
    Zhang, Shigeng
    Wang, Weiping
    Zhang, Jian
    Liu, Xuan
    ALGORITHMS AND ARCHITECTURES FOR PARALLEL PROCESSING, ICA3PP 2023, PT V, 2024, 14491 : 238 - 256
  • [36] A study of natural robustness of deep reinforcement learning algorithms towards adversarial perturbations
    Liu, Qisai
    Lee, Xian Yeow
    Sarkar, Soumik
    AI OPEN, 2024, 5 : 126 - 141
  • [37] ENHANCING ADVERSARIAL ROBUSTNESS FOR IMAGE CLASSIFICATION BY REGULARIZING CLASS LEVEL FEATURE DISTRIBUTION
    Yu, Cheng
    Xue, Youze
    Chen, Jiansheng
    Wang, Yu
    Ma, Huimin
    2021 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2021, : 494 - 498
  • [38] Enhancing the adversarial robustness in medical image classification: exploring adversarial machine learning with vision transformers-based models
    Elif Kanca Gulsoy
    Selen Ayas
    Elif Baykal Kablan
    Murat Ekinci
    Neural Computing and Applications, 2025, 37 (12) : 7971 - 7989
  • [39] Between-Class Adversarial Training for Improving Adversarial Robustness of Image Classification
    Wang, Desheng
    Jin, Weidong
    Wu, Yunpu
    SENSORS, 2023, 23 (06)
  • [40] Improving accuracy and robustness of deep convolutional neural network based thoracic OAR segmentation
    Feng, Xue
    Bernard, Mark E.
    Hunter, Thomas
    Chen, Quan
    PHYSICS IN MEDICINE AND BIOLOGY, 2020, 65 (07)