Exploring adversarial examples and adversarial robustness of convolutional neural networks by mutual information

被引:0
|
作者
Zhang J. [1 ]
Qian W. [1 ]
Cao J. [2 ,3 ]
Xu D. [1 ]
机构
[1] School of Information Science and Engineering, Yunnan University, Kunming
[2] School of Mathematics, Southeast University, Nanjing
[3] Ahlia University, Manama
基金
中国国家自然科学基金;
关键词
Adversarial attacks; Adversarial examples; Deep neural networks; Mutual information;
D O I
10.1007/s00521-024-09774-z
中图分类号
学科分类号
摘要
Convolutional neural networks (CNNs) are susceptible to adversarial examples, which are similar to original examples but contain malicious perturbations. Adversarial training is a simple and effective defense method to improve the robustness of CNNs to adversarial examples. Many works explore the mechanism behind adversarial examples and adversarial training. However, mutual information is rarely present in the interpretation of these counter-intuitive phenomena. This work investigates similarities and differences between normally trained CNNs (NT-CNNs) and adversarially trained CNNs (AT-CNNs) from the mutual information perspective. We show that although mutual information trends of NT-CNNs and AT-CNNs are similar throughout training for original and adversarial examples, there exists an obvious difference. Compared with NT-CNNs, AT-CNNs achieve a lower clean accuracy and extract less information from the input. CNNs trained with different methods have different preferences for certain types of information; NT-CNNs tend to extract texture-based information from the input, while AT-CNNs prefer shape-based information. The reason why adversarial examples mislead CNNs may be that they contain more texture-based information about other classes. Furthermore, we also analyze the mutual information estimators used in this work and find that they outline the geometric properties of the middle layer’s output. © The Author(s), under exclusive licence to Springer-Verlag London Ltd., part of Springer Nature 2024.
引用
收藏
页码:14379 / 14394
页数:15
相关论文
共 50 条
  • [21] Adversarial robustness improvement for deep neural networks
    Charis Eleftheriadis
    Andreas Symeonidis
    Panagiotis Katsaros
    Machine Vision and Applications, 2024, 35
  • [22] On the Robustness of Bayesian Neural Networks to Adversarial Attacks
    Bortolussi, Luca
    Carbone, Ginevra
    Laurenti, Luca
    Patane, Andrea
    Sanguinetti, Guido
    Wicker, Matthew
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, : 1 - 14
  • [23] Understanding adversarial robustness against on-manifold adversarial examples
    Xiao, Jiancong
    Yang, Liusha
    Fan, Yanbo
    Wang, Jue
    Luo, Zhi-Quan
    PATTERN RECOGNITION, 2025, 159
  • [24] Convolutional and generative adversarial neural networks in manufacturing
    Kusiak, Andrew
    INTERNATIONAL JOURNAL OF PRODUCTION RESEARCH, 2020, 58 (05) : 1594 - 1604
  • [25] Characterizing Adversarial Samples of Convolutional Neural Networks
    Jiang, Cheng
    Zhao, Qiyang
    Liu, Yuzhong
    2018 11TH INTERNATIONAL CONGRESS ON IMAGE AND SIGNAL PROCESSING, BIOMEDICAL ENGINEERING AND INFORMATICS (CISP-BMEI 2018), 2018,
  • [26] A Hybrid Bayesian-Convolutional Neural Network for Adversarial Robustness
    Khong, Thi Thu Thao
    Nakada, Takashi
    Nakashima, Yasuhiko
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2022, E105D (07) : 1308 - 1319
  • [27] ARGAN: Adversarially Robust Generative Adversarial Networks for Deep Neural Networks Against Adversarial Examples
    Choi, Seok-Hwan
    Shin, Jin-Myeong
    Liu, Peng
    Choi, Yoon-Ho
    IEEE ACCESS, 2022, 10 : 33602 - 33615
  • [28] ARGAN: Adversarially Robust Generative Adversarial Networks for Deep Neural Networks Against Adversarial Examples
    Choi, Seok-Hwan
    Shin, Jin-Myeong
    Liu, Peng
    Choi, Yoon-Ho
    IEEE Access, 2022, 10 : 33602 - 33615
  • [29] On the robustness of randomized classifiers to adversarial examples
    Rafael Pinot
    Laurent Meunier
    Florian Yger
    Cédric Gouy-Pailler
    Yann Chevaleyre
    Jamal Atif
    Machine Learning, 2022, 111 : 3425 - 3457
  • [30] On the Robustness of Vision Transformers to Adversarial Examples
    Mahmood, Kaleel
    Mahmood, Rigel
    van Dijk, Marten
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 7818 - 7827