Flexible Bayesian Inference by Weight Transfer for Robust Deep Neural Networks

被引:1
作者
Thi Thu Thao Khong [1 ]
Nakada, Takashi [1 ]
Nakashima, Yasuhiko [1 ]
机构
[1] Nara Inst Sci & Technol, Ikoma 6300192, Japan
关键词
Deep Neural Network; Bayesian Neural Network; image classification; adversarial attacks; adversarial training;
D O I
10.1587/transinf.2021EDP7046
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Adversarial attacks are viewed as a danger to Deep Neural Networks (DNNs), which reveal a weakness of deep learning models in security-critical applications. Recent findings have been presented adversarial training as an outstanding defense method against adversaries. Nonetheless, adversarial training is a challenge with respect to big datasets and large networks. It is believed that, unless making DNN architectures larger, DNNs would be hard to strengthen the robustness to adversarial examples. In order to avoid iteratively adversarial training, our algorithm is Bayes without Bayesian Learning (BwoBL) that performs the ensemble inference to improve the robustness. As an application of transfer learning, we use learned parameters of pretrained DNNs to build Bayesian Neural Networks (BNNs) and focus on Bayesian inference without costing Bayesian learning. In comparison with no adversarial training, our method is more robust than activation functions designed to enhance adversarial robustness. Moreover, BwoBL can easily integrate into any pretrained DNN, not only Convolutional Neural Networks (CNNs) but also other DNNs, such as Self-Attention Networks (SANs) that outperform convolutional counterparts. BwoBL is also convenient to apply to scaling networks, e.g., ResNet and EfficientNet, with better performance. Especially, our algorithm employs a variety of DNN architectures to construct BNNs against a diversity of adversarial attacks on a large-scale dataset. In particular, under l(infinity )norm PGD attack of pixel perturbation epsilon = 4/255 with 100 iterations on ImageNet, our proposal in ResNets, SANs, and EfficientNets increase by 58.18% top-5 accuracy on average, which are combined with naturally pretrained ResNets, SANs, and EfficientNets. This enhancement is 62.26% on average below l(2) norm C&W attack. The combination of our proposed method with pretrained EfficientNets on both natural and adversarial images (EfficientNet-ADV) drastically boosts the robustness resisting PGD and C&W attacks without additional training. Our EfficientNet-ADV-B7 achieves the cutting-edge top-5 accuracy, which is 92.14% and 94.20% on adversarial ImageNet generated by powerful PGD and C&W attacks, respectively.
引用
收藏
页码:1981 / 1991
页数:11
相关论文
共 41 条
  • [1] [Anonymous], 2017, AISEC
  • [2] [Anonymous], 2017, P ACM WORKSH ART INT, DOI DOI 10.1145/3128572.3140449
  • [3] Barber D., 1998, Neural Networks and Machine Learning. Proceedings, P215
  • [4] Bishop C., 1995, Bayesian methods for neural networks
  • [5] Blundell C, 2015, PR MACH LEARN RES, V37, P1613
  • [6] Towards Evaluating the Robustness of Neural Networks
    Carlini, Nicholas
    Wagner, David
    [J]. 2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2017, : 39 - 57
  • [7] Dhillon G. S., 2018, P ICLR
  • [8] Ding G. W., 2019, arXiv preprint, arXiv:1902.07623
  • [9] Boosting Adversarial Attacks with Momentum
    Dong, Yinpeng
    Liao, Fangzhou
    Pang, Tianyu
    Su, Hang
    Zhu, Jun
    Hu, Xiaolin
    Li, Jianguo
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 9185 - 9193
  • [10] Gal Y., 2015, Bayesian convolutional neural networks with bernoulli approximate variational inference