Exploring adversarial examples and adversarial robustness of convolutional neural networks by mutual information

被引:0
|
作者
Zhang J. [1 ]
Qian W. [1 ]
Cao J. [2 ,3 ]
Xu D. [1 ]
机构
[1] School of Information Science and Engineering, Yunnan University, Kunming
[2] School of Mathematics, Southeast University, Nanjing
[3] Ahlia University, Manama
基金
中国国家自然科学基金;
关键词
Adversarial attacks; Adversarial examples; Deep neural networks; Mutual information;
D O I
10.1007/s00521-024-09774-z
中图分类号
学科分类号
摘要
Convolutional neural networks (CNNs) are susceptible to adversarial examples, which are similar to original examples but contain malicious perturbations. Adversarial training is a simple and effective defense method to improve the robustness of CNNs to adversarial examples. Many works explore the mechanism behind adversarial examples and adversarial training. However, mutual information is rarely present in the interpretation of these counter-intuitive phenomena. This work investigates similarities and differences between normally trained CNNs (NT-CNNs) and adversarially trained CNNs (AT-CNNs) from the mutual information perspective. We show that although mutual information trends of NT-CNNs and AT-CNNs are similar throughout training for original and adversarial examples, there exists an obvious difference. Compared with NT-CNNs, AT-CNNs achieve a lower clean accuracy and extract less information from the input. CNNs trained with different methods have different preferences for certain types of information; NT-CNNs tend to extract texture-based information from the input, while AT-CNNs prefer shape-based information. The reason why adversarial examples mislead CNNs may be that they contain more texture-based information about other classes. Furthermore, we also analyze the mutual information estimators used in this work and find that they outline the geometric properties of the middle layer’s output. © The Author(s), under exclusive licence to Springer-Verlag London Ltd., part of Springer Nature 2024.
引用
收藏
页码:14379 / 14394
页数:15
相关论文
共 50 条
  • [21] Spot evasion attacks: Adversarial examples for license plate recognition systems with convolutional neural networks
    Qian, Yaguan
    Ma, Danfeng
    Wang, Bin
    Pan, Jun
    Wang, Jiamin
    Gu, Zhaoquan
    Chen, Jianhai
    Zhou, Wujie
    Lei, Jingsheng
    COMPUTERS & SECURITY, 2020, 95 (95)
  • [22] Adversarial Robustness of Neural Networks from the Perspective of Lipschitz Calculus: A Survey
    Zuehlke, Monty-maximilian
    Kudenko, Daniel
    ACM COMPUTING SURVEYS, 2025, 57 (06)
  • [23] Exploring the Adversarial Frontier: Quantifying Robustness via Adversarial Hypervolume
    Guo, Ping
    Gong, Cheng
    Lin, Xi
    Yang, Zhiyuan
    Zhang, Qingfu
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2025, 9 (02): : 1367 - 1378
  • [24] Adversarial Attacks in Modulation Recognition With Convolutional Neural Networks
    Lin, Yun
    Zhao, Haojun
    Ma, Xuefei
    Tu, Ya
    Wang, Meiyu
    IEEE TRANSACTIONS ON RELIABILITY, 2021, 70 (01) : 389 - 401
  • [25] On the Relationship between Generalization and Robustness to Adversarial Examples
    Pedraza, Anibal
    Deniz, Oscar
    Bueno, Gloria
    SYMMETRY-BASEL, 2021, 13 (05):
  • [26] Robustness to adversarial examples can be improved with overfitting
    Deniz, Oscar
    Pedraza, Anibal
    Vallez, Noelia
    Salido, Jesus
    Bueno, Gloria
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2020, 11 (04) : 935 - 944
  • [27] Regularizing Hard Examples Improves Adversarial Robustness
    Lee, Hyungyu
    Lee, Saehyung
    Bae, Ho
    Yoon, Sungroh
    JOURNAL OF MACHINE LEARNING RESEARCH, 2025, 26
  • [28] Robustness to adversarial examples can be improved with overfitting
    Oscar Deniz
    Anibal Pedraza
    Noelia Vallez
    Jesus Salido
    Gloria Bueno
    International Journal of Machine Learning and Cybernetics, 2020, 11 : 935 - 944
  • [29] Deep neural rejection against adversarial examples
    Sotgiu, Angelo
    Demontis, Ambra
    Melis, Marco
    Biggio, Battista
    Fumera, Giorgio
    Feng, Xiaoyi
    Roli, Fabio
    EURASIP JOURNAL ON INFORMATION SECURITY, 2020, 2020 (01)
  • [30] Training Neural Networks with Random Noise Images for Adversarial Robustness
    Park, Ji-Young
    Liu, Lin
    Li, Jiuyong
    Liu, Jixue
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 3358 - 3362