Adversarial Robustness with Partial Isometry

被引:2
作者
Shi-Garrier, Loic [1 ]
Bouaynaya, Nidhal Carla [2 ]
Delahaye, Daniel [1 ]
机构
[1] Univ Toulouse, ENAC, F-31400 Toulouse, France
[2] Rowan Univ, Dept Elect & Comp Engn, Glassboro, NJ 08028 USA
关键词
adversarial robustness; information geometry; fisher information metric; multi-class classification;
D O I
10.3390/e26020103
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
Despite their remarkable performance, deep learning models still lack robustness guarantees, particularly in the presence of adversarial examples. This significant vulnerability raises concerns about their trustworthiness and hinders their deployment in critical domains that require certified levels of robustness. In this paper, we introduce an information geometric framework to establish precise robustness criteria for l2 white-box attacks in a multi-class classification setting. We endow the output space with the Fisher information metric and derive criteria on the input-output Jacobian to ensure robustness. We show that model robustness can be achieved by constraining the model to be partially isometric around the training points. We evaluate our approach using MNIST and CIFAR-10 datasets against adversarial attacks, revealing its substantial improvements over defensive distillation and Jacobian regularization for medium-sized perturbations and its superior robustness performance to adversarial training for large perturbations, all while maintaining the desired accuracy.
引用
收藏
页数:18
相关论文
共 50 条
  • [31] On the Robustness of Bayesian Neural Networks to Adversarial Attacks
    Bortolussi, Luca
    Carbone, Ginevra
    Laurenti, Luca
    Patane, Andrea
    Sanguinetti, Guido
    Wicker, Matthew
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, : 1 - 14
  • [32] On the adversarial robustness of generative autoencoders in the latent space
    Lu, Mingfei
    Chen, Badong
    NEURAL COMPUTING & APPLICATIONS, 2024, : 8109 - 8123
  • [33] IDEA: Invariant defense for graph adversarial robustness
    Tao, Shuchang
    Cao, Qi
    Shen, Huawei
    Wu, Yunfan
    Xu, Bingbing
    Cheng, Xueqi
    INFORMATION SCIENCES, 2024, 680
  • [34] Evaluating the transferability of adversarial robustness to target domains
    Kopetzki, Anna-Kathrin
    Bojchevski, Aleksandar
    Guennemann, Stephan
    KNOWLEDGE AND INFORMATION SYSTEMS, 2025, : 4139 - 4206
  • [35] Regularizing Hard Examples Improves Adversarial Robustness
    Lee, Hyungyu
    Lee, Saehyung
    Bae, Ho
    Yoon, Sungroh
    JOURNAL OF MACHINE LEARNING RESEARCH, 2025, 26
  • [36] Adversarial robustness improvement for deep neural networks
    Charis Eleftheriadis
    Andreas Symeonidis
    Panagiotis Katsaros
    Machine Vision and Applications, 2024, 35
  • [37] Enhancing adversarial robustness with randomized interlayer processing
    Mohammed, Ameer
    Ali, Ziad
    Ahmad, Imtiaz
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 245
  • [38] Adversarial robustness improvement for deep neural networks
    Eleftheriadis, Charis
    Symeonidis, Andreas
    Katsaros, Panagiotis
    MACHINE VISION AND APPLICATIONS, 2024, 35 (03)
  • [39] Delving into Adversarial Robustness on Document Tampering Localization
    Shao, Huiru
    Qian, Zhuang
    Huang, Kaizhu
    Wang, Wei
    Huang, Xiaowei
    Wang, Qiufeng
    COMPUTER VISION - ECCV 2024, PT LXV, 2025, 15123 : 290 - 306
  • [40] Improving Adversarial Robustness by Reconstructing Interclass Relationships
    Xu, Li
    Guo, Huiting
    Yang, Zejin
    Wan, Xu
    Fan, Chunlong
    PROCEEDINGS OF THE 2024 27 TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN, CSCWD 2024, 2024, : 1968 - 1973