Measuring the Effect of Causal Disentanglement on the Adversarial Robustness of Neural Network Models

被引:0
作者
Ness, Preben M. [1 ]
Marijan, Dusica [1 ]
Bose, Sunanda [1 ]
机构
[1] Simula Res Lab, Oslo, Norway
来源
PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023 | 2023年
关键词
causal inference; causal neural network; neural networks; adversarial robustness; representation learning; computer vision;
D O I
10.1145/3583780.3614960
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Causal Neural Network models have shown high levels of robustness to adversarial attacks as well as an increased capacity for generalisation tasks such as few-shot learning and rare-context classification compared to traditional Neural Networks. This robustness is argued to stem from the disentanglement of causal and confounder input signals. However, no quantitative study has yet measured the level of disentanglement achieved by these types of causal models or assessed how this relates to their adversarial robustness. Existing causal disentanglement metrics are not applicable to deterministic models trained on real-world datasets. We, therefore, utilise metrics of content/style disentanglement from the field of Computer Vision to measure different aspects of the causal disentanglement for four state-of-the-art causal Neural Network models. By re-implementing these models with a common ResNet18 architecture we are able to fairly measure their adversarial robustness on three standard image classification benchmarking datasets under seven common white-box attacks. We find a strong association (r=0.820, p=0.001) between the degree to which models decorrelate causal and confounder signals and their adversarial robustness. Additionally, we find a moderate negative association between the pixel-level information content of the confounder signal and adversarial robustness (r=-0.597, p=0.040).
引用
收藏
页码:1907 / 1916
页数:10
相关论文
共 40 条
  • [1] Amos David, 2018, THEORETICAL PHYS DEE
  • [2] Representation Learning: A Review and New Perspectives
    Bengio, Yoshua
    Courville, Aaron
    Vincent, Pascal
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2013, 35 (08) : 1798 - 1828
  • [3] Measuring Disentanglement: A Review of Metrics
    Carbonneau, Marc-Andre
    Zaidi, Julian
    Boilard, Jonathan
    Gagnon, Ghyslain
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (07) : 8747 - 8761
  • [4] Towards Evaluating the Robustness of Neural Networks
    Carlini, Nicholas
    Wagner, David
    [J]. 2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2017, : 39 - 57
  • [5] Chalupka K, 2015, UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, P181
  • [6] Do Kien, 2019, INT C LEARN REPR
  • [7] Duan Sunny, 2019, INT C LEARN REPR
  • [8] Ferraro Stefano, 2022, ARXIV220909097
  • [9] Goldfeld Z., 2019, 36 INT C MACH LEARN, P4153
  • [10] Goodfellow I., 2015, P INT C LEARN REPR