Is Robustness the Cost of Accuracy? - A Comprehensive Study on the Robustness of 18 Deep Image Classification Models

被引:183
|
作者
Su, Dong [1 ]
Zhang, Huan [2 ]
Chen, Hongge [3 ]
Yi, Jinfeng [4 ]
Chen, Pin-Yu [1 ]
Gao, Yupeng [1 ]
机构
[1] IBM Res, New York, NY 10598 USA
[2] Univ Calif Davis, Davis, CA 95616 USA
[3] MIT, Cambridge, MA 02139 USA
[4] JD AI Res, Beijing, Peoples R China
来源
COMPUTER VISION - ECCV 2018, PT XII | 2018年 / 11216卷
关键词
Deep neural networks; Adversarial attacks; Robustness;
D O I
10.1007/978-3-030-01258-8_39
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The prediction accuracy has been the long-lasting and sole standard for comparing the performance of different image classification models, including the ImageNet competition. However, recent studies have highlighted the lack of robustness in well-trained deep neural networks to adversarial examples. Visually imperceptible perturbations to natural images can easily be crafted and mislead the image classifiers towards misclassification. To demystify the trade-offs between robustness and accuracy, in this paper we thoroughly benchmark 18 ImageNet models using multiple robustness metrics, including the distortion, success rate and transferability of adversarial examples between 306 pairs of models. Our extensive experimental results reveal several new insights: (1) linear scaling law - the empirical l(2) and l(infinity) distortion metrics scale linearly with the logarithm of classification error; (2) model architecture is a more critical factor to robustness than model size, and the disclosed accuracy-robustness Pareto frontier can be used as an evaluation criterion for ImageNet model designers; (3) for a similar network architecture, increasing network depth slightly improves robustness in l(infinity) distortion; (4) there exist models (in VGG family) that exhibit high adversarial transferability, while most adversarial examples crafted from one model can only be transferred within the same family. Experiment code is publicly available at https://github.com/huanzhang12/Adversarial_Survey.
引用
收藏
页码:644 / 661
页数:18
相关论文
共 50 条
  • [11] ROBUSTNESS OF DEEP CONVOLUTIONAL NEURAL NETWORKS FOR IMAGE DEGRADATIONS
    Ghosh, Sanjukta
    Shet, Rohan
    Amon, Peter
    Hutter, Andreas
    Kaup, Andre
    2018 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2018, : 2916 - 2920
  • [12] Active dropblock: Method to enhance deep model accuracy and robustness
    Yao, Jie
    Xing, Weiwei
    Wang, Dongdong
    Xing, Jintao
    Wang, Liqiang
    NEUROCOMPUTING, 2021, 454 : 189 - 200
  • [13] Predicted Robustness as QoS for Deep Neural Network Models
    Wang, Yue-Huan
    Li, Ze-Nan
    Xu, Jing-Wei
    Yu, Ping
    Chen, Taolue
    Ma, Xiao-Xing
    JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY, 2020, 35 (05) : 999 - 1015
  • [14] Predicted Robustness as QoS for Deep Neural Network Models
    Yue-Huan Wang
    Ze-Nan Li
    Jing-Wei Xu
    Ping Yu
    Taolue Chen
    Xiao-Xing Ma
    Journal of Computer Science and Technology, 2020, 35 : 999 - 1015
  • [15] A Fast Robustness Quantification Method for Evaluating Typical Deep Learning Models by Generally Image Processing
    Li, Haocong
    Cheng, Yunjia
    Ren, Wei
    Zhu, Tianqing
    2020 IEEE INTL SYMP ON PARALLEL & DISTRIBUTED PROCESSING WITH APPLICATIONS, INTL CONF ON BIG DATA & CLOUD COMPUTING, INTL SYMP SOCIAL COMPUTING & NETWORKING, INTL CONF ON SUSTAINABLE COMPUTING & COMMUNICATIONS (ISPA/BDCLOUD/SOCIALCOM/SUSTAINCOM 2020), 2020, : 110 - 117
  • [16] SPLAS: An Autonomous Lightweight and Versatile Image Classification Robustness Benchmark
    Bian, Yiming
    Somani, Arun K.
    IEEE ACCESS, 2025, 13 : 51848 - 51861
  • [17] Comparing the Robustness of Classical and Deep Learning Techniques for Text Classification
    Quynh Tran
    Shpileuskaya, Krystsina
    Zaunseder, Elaine
    Putzar, Larissa
    Blankenburg, Sven
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [18] Evaluating and Improving Adversarial Robustness of Deep Learning Models for Intelligent Vehicle Safety
    Hussain, Manzoor
    Hong, Jang-Eui
    IEEE TRANSACTIONS ON RELIABILITY, 2024,
  • [19] Improving robustness of a deep learning-based lung-nodule classification model of CT images with respect to image noise
    Gao, Yin
    Xiong, Jennifer
    Shen, Chenyang
    Jia, Xun
    PHYSICS IN MEDICINE AND BIOLOGY, 2021, 66 (24)
  • [20] Robustness evaluation of image classification models based on edge features: Tight and Non-Tight boundaries
    Lu, Hui
    Mei, Ting
    Wang, Shiqi
    Zhang, Ruoliu
    Mao, Kefei
    NEUROCOMPUTING, 2025, 623