Improving adversarial robustness of Bayesian neural networks via multi-task adversarial training

被引:9
|
作者
Chen, Xu [1 ]
Liu, Chuancai [1 ]
Zhao, Yue [2 ]
Jia, Zhiyang [3 ]
Jin, Ge [1 ]
机构
[1] Nanjing Univ Sci & Technol, Sch Comp Sci & Engn, Nanjing 210094, Peoples R China
[2] Yunnan Univ, Inst Math & Stat, Kunming 650091, Yunnan, Peoples R China
[3] China Univ Petr Beijing Karamay, Dept Comp Sci, Karamay 834000, Peoples R China
关键词
Bayesian neural networks; Adversarial training; Variational inference; Multi-task loss; Adversarial robustness;
D O I
10.1016/j.ins.2022.01.051
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Bayesian neural networks (BNNs) are used in many tasks because they provide a probabilistic representation of deep learning models by placing a distribution over the model parameters. Although BNNs are a more robust deep learning paradigm than vanilla deep neural networks, their ability to handle adversarial attacks in practice remains limited. In this study, we propose a novel multi-task adversarial training approach for improving the adversarial robustness of BNNs. Specifically, we first generate diverse and stronger adversarial examples for adversarial training by maximising a multi-task loss. This multi-task loss is a combination of the unsupervised feature scattering loss and supervised margin loss. Then, we find the model parameters by minimising another multi-task loss composed of the feature loss and variational inference loss. The feature loss is defined based on distance parallel to l parallel to(p), which measures the difference between the two feature representations extracted from the clean and adversarial examples. Minimising the feature loss improves the feature similarity and helps the model learn more robust features, resulting in enhanced robustness. Extensive experiments are conducted on four benchmark datasets in white-box and black-box attack scenarios. The experimental results demonstrate that the proposed approach significantly improves the adversarial robustness compared with several state-of-the-art defence methods. (C) 2022 Elsevier Inc. All rights reserved.
引用
收藏
页码:156 / 173
页数:18
相关论文
共 50 条
  • [1] Improving the Robustness of Deep Neural Networks via Adversarial Training with Triplet Loss
    Li, Pengcheng
    Yi, Jinfeng
    Zhou, Bowen
    Zhang, Lijun
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 2909 - 2915
  • [2] Defending Against Adversarial Attack Towards Deep Neural Networks Via Collaborative Multi-Task Training
    Wang, Derui
    Li, Chaoran
    Wen, Sheng
    Nepal, Surya
    Xiang, Yang
    IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, 2022, 19 (02) : 953 - 965
  • [3] Adversarial Robustness Certification for Bayesian Neural Networks
    Wicker, Matthew
    Platzer, Andre
    Laurenti, Luca
    Kwiatkowska, Marta
    FORMAL METHODS, PT I, FM 2024, 2025, 14933 : 3 - 28
  • [4] On the Robustness of Bayesian Neural Networks to Adversarial Attacks
    Bortolussi, Luca
    Carbone, Ginevra
    Laurenti, Luca
    Patane, Andrea
    Sanguinetti, Guido
    Wicker, Matthew
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, : 1 - 14
  • [5] Adversarial Robustness in Multi-Task Learning: Promises and Illusions
    Ghamizi, Salah
    Cordy, Maxime
    Papadakis, Mike
    Le Traon, Yves
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 697 - 705
  • [6] Improving Adversarial Robustness of Deep Neural Networks via Linear Programming
    Tang, Xiaochao
    Yang, Zhengfeng
    Fu, Xuanming
    Wang, Jianlin
    Zeng, Zhenbing
    THEORETICAL ASPECTS OF SOFTWARE ENGINEERING, TASE 2022, 2022, 13299 : 326 - 343
  • [7] Improving Bayesian Neural Networks by Adversarial Sampling
    Zhang, Jiaru
    Hua, Yang
    Song, Tao
    Wang, Hao
    Xue, Zhengui
    Ma, Ruhui
    Guan, Haibing
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 10110 - 10117
  • [8] Multi-Task Learning With Self-Defined Tasks for Adversarial Robustness of Deep Networks
    Hyun, Changhun
    Park, Hyeyoung
    IEEE ACCESS, 2024, 12 : 83248 - 83259
  • [9] An orthogonal classifier for improving the adversarial robustness of neural networks
    Xu, Cong
    Li, Xiang
    Yang, Min
    INFORMATION SCIENCES, 2022, 591 : 251 - 262
  • [10] Sliced Wasserstein adversarial training for improving adversarial robustness
    Lee W.
    Lee S.
    Kim H.
    Lee J.
    Journal of Ambient Intelligence and Humanized Computing, 2024, 15 (08) : 3229 - 3242