Is Robustness the Cost of Accuracy? - A Comprehensive Study on the Robustness of 18 Deep Image Classification Models

被引:183
|
作者
Su, Dong [1 ]
Zhang, Huan [2 ]
Chen, Hongge [3 ]
Yi, Jinfeng [4 ]
Chen, Pin-Yu [1 ]
Gao, Yupeng [1 ]
机构
[1] IBM Res, New York, NY 10598 USA
[2] Univ Calif Davis, Davis, CA 95616 USA
[3] MIT, Cambridge, MA 02139 USA
[4] JD AI Res, Beijing, Peoples R China
来源
COMPUTER VISION - ECCV 2018, PT XII | 2018年 / 11216卷
关键词
Deep neural networks; Adversarial attacks; Robustness;
D O I
10.1007/978-3-030-01258-8_39
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The prediction accuracy has been the long-lasting and sole standard for comparing the performance of different image classification models, including the ImageNet competition. However, recent studies have highlighted the lack of robustness in well-trained deep neural networks to adversarial examples. Visually imperceptible perturbations to natural images can easily be crafted and mislead the image classifiers towards misclassification. To demystify the trade-offs between robustness and accuracy, in this paper we thoroughly benchmark 18 ImageNet models using multiple robustness metrics, including the distortion, success rate and transferability of adversarial examples between 306 pairs of models. Our extensive experimental results reveal several new insights: (1) linear scaling law - the empirical l(2) and l(infinity) distortion metrics scale linearly with the logarithm of classification error; (2) model architecture is a more critical factor to robustness than model size, and the disclosed accuracy-robustness Pareto frontier can be used as an evaluation criterion for ImageNet model designers; (3) for a similar network architecture, increasing network depth slightly improves robustness in l(infinity) distortion; (4) there exist models (in VGG family) that exhibit high adversarial transferability, while most adversarial examples crafted from one model can only be transferred within the same family. Experiment code is publicly available at https://github.com/huanzhang12/Adversarial_Survey.
引用
收藏
页码:644 / 661
页数:18
相关论文
共 50 条
  • [21] Towards evaluating the robustness of deep diagnostic models by adversarial attack
    Xu, Mengting
    Zhang, Tao
    Li, Zhongnian
    Liu, Mingxia
    Zhang, Daoqiang
    MEDICAL IMAGE ANALYSIS, 2021, 69
  • [22] Reducing Textural Bias Improves Robustness of Deep Segmentation Models
    Chai, Seoin
    Rueckert, Daniel
    Fetit, Ahmed E.
    MEDICAL IMAGE UNDERSTANDING AND ANALYSIS (MIUA 2021), 2021, 12722 : 294 - 304
  • [23] Robustness of Deep Learning Models in Dermatological Evaluation: A Critical Assessment
    Mishra, Sourav
    Chaudhury, Subhajit
    Imaizumi, Hideaki
    Yamasaki, Toshihiko
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2021, E104D (03) : 419 - 429
  • [24] Improving the robustness and accuracy of biomedical language models through adversarial training
    Moradi, Milad
    Samwald, Matthias
    JOURNAL OF BIOMEDICAL INFORMATICS, 2022, 132
  • [25] Triangular Trade-off between Robustness, Accuracy, and Fairness in Deep Neural Networks: A Survey
    Li, Jingyang
    Li, Guoqiang
    ACM COMPUTING SURVEYS, 2025, 57 (06)
  • [26] RobustRMC: Robustness Interpretable Deep Neural Network for Radio Modulation Classification
    Chen, Jinyin
    Liao, Danxin
    Zheng, Shilian
    Ye, Linhui
    Jia, Chenyu
    Zheng, Haibin
    Xiang, Sheng
    IEEE TRANSACTIONS ON COGNITIVE COMMUNICATIONS AND NETWORKING, 2024, 10 (04) : 1218 - 1240
  • [27] Robustness of the directional neighborhoods approach to image classification with respect to spatial correlation
    Dagne, GA
    Press, SJ
    AMERICAN STATISTICAL ASSOCIATION - 1996 PROCEEDINGS OF THE SECTION ON BAYESIAN STATISTICAL SCIENCE, 1996, : 64 - 69
  • [28] On the robustness of deep learning-based lung-nodule classification for CT images with respect to image noise
    Shen, Chenyang
    Tsai, Min-Yu
    Chen, Liyuan
    Li, Shulong
    Nguyen, Dan
    Wang, Jing
    Jiang, Steve B.
    Jia, Xun
    PHYSICS IN MEDICINE AND BIOLOGY, 2020, 65 (24)
  • [29] From modern CNNs to vision transformers: Assessing the performance, robustness, and classification strategies of deep learning models in histopathology
    Springenberg, Maximilian
    Frommholz, Annika
    Wenzel, Markus
    Weicken, Eva
    Ma, Jackie
    Strodthoff, Nils
    MEDICAL IMAGE ANALYSIS, 2023, 87
  • [30] A COMPARATIVE STUDY OF ROBUSTNESS OF DEEP LEARNING APPROACHES FOR VAD
    Tong, Sibo
    Gu, Hao
    Yu, Kai
    2016 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING PROCEEDINGS, 2016, : 5695 - 5699