Feature Distillation in Deep Attention Network Against Adversarial Examples

被引:7
|
作者
Chen, Xin [1 ,2 ,3 ,4 ]
Weng, Jian [5 ]
Deng, Xiaoling [1 ,2 ,3 ,4 ]
Luo, Weiqi [5 ]
Lan, Yubin [1 ,2 ,3 ,4 ]
Tian, Qi [6 ]
机构
[1] South China Agr Univ, Coll Elect Engn, Coll Artificial Intelligence, Guangzhou 510642, Peoples R China
[2] Natl Ctr Int Collaborat Res Precis Agr Aviat Pest, Guangzhou 510642, Peoples R China
[3] Guangdong Lab Lingnan Modern Agr, Guangzhou 510642, Peoples R China
[4] Guangdong Engn Technol Res Ctr Smart Agr, Guangzhou 510642, Peoples R China
[5] Jinan Univ, Coll Informat Sci & Technol, Guangzhou 510632, Peoples R China
[6] Huawei Cloud & AI, Shenzhen 518100, Peoples R China
基金
中国国家自然科学基金;
关键词
Perturbation methods; Robustness; Image coding; Training; Neural networks; Task analysis; Research and development; Adversarial examples; deep attention mechanism; frequency decomposition; image classification;
D O I
10.1109/TNNLS.2021.3113342
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks (DNNs) are easily fooled by adversarial examples. Most existing defense strategies defend against adversarial examples based on full information of whole images. In reality, one possible reason as to why humans are not sensitive to adversarial perturbations is that the human visual mechanism often concentrates on most important regions of images. A deep attention mechanism has been applied in many computer fields and has achieved great success. Attention modules are composed of an attention branch and a trunk branch. The encoder/decoder architecture in the attention branch has potential of compressing adversarial perturbations. In this article, we theoretically prove that attention modules can compress adversarial perturbations by destroying potential linear characteristics of DNNs. Considering the distribution characteristics of adversarial perturbations in different frequency bands, we design and compare three types of attention modules based on frequency decomposition and reorganization to defend against adversarial examples. Moreover, we find that our designed attention modules can obtain high classification accuracies on clean images by locating attention regions more accurately. Experimental results on the CIFAR and ImageNet dataset demonstrate that frequency reorganization in attention modules can not only achieve good robustness to adversarial perturbations, but also obtain comparable, even higher classification, accuracies on clean images. Moreover, our proposed attention modules can be integrated with existing defense strategies as components to further improve adversarial robustness.
引用
收藏
页码:3691 / 3705
页数:15
相关论文
共 50 条
  • [31] Enhancing Robustness Against Adversarial Examples in Network Intrusion Detection Systems
    Hashemi, Mohammad J.
    Keller, Eric
    2020 IEEE CONFERENCE ON NETWORK FUNCTION VIRTUALIZATION AND SOFTWARE DEFINED NETWORKS (NFV-SDN), 2020, : 37 - 43
  • [32] Class Incremental Learning With Deep Contrastive Learning and Attention Distillation
    Zhu, Jitao
    Luo, Guibo
    Duan, Baishan
    Zhu, Yuesheng
    IEEE SIGNAL PROCESSING LETTERS, 2024, 31 : 1224 - 1228
  • [33] Resisting Deep Learning Models Against Adversarial Attack Transferability via Feature Randomization
    Nowroozi, Ehsan
    Mohammadi, Mohammadreza
    Golmohammadi, Pargol
    Mekdad, Yassine
    Conti, Mauro
    Uluagac, Selcuk
    IEEE TRANSACTIONS ON SERVICES COMPUTING, 2024, 17 (01) : 18 - 29
  • [34] Regional attention generative adversarial network
    Wang, Wei
    Hu, Haifeng
    Huang, Yi
    Ruan, Chongchong
    Chen, Dihu
    ELECTRONICS LETTERS, 2019, 55 (08) : 459 - +
  • [35] Defed: An Edge-Feature-Enhanced Image Denoised Network Against Adversarial Attacks for Secure Internet of Things
    Xiao, Yang
    Yan, Chengjia
    Lyu, Shuo
    Pei, Qingqi
    Liu, Ximeng
    Zhang, Ning
    Dong, Mianxiong
    IEEE INTERNET OF THINGS JOURNAL, 2023, 10 (08) : 6836 - 6848
  • [36] Adversarial Minimax Training for Robustness Against Adversarial Examples
    Komiyama, Ryota
    Hattori, Motonobu
    NEURAL INFORMATION PROCESSING (ICONIP 2018), PT II, 2018, 11302 : 690 - 699
  • [37] DDSA: A Defense Against Adversarial Attacks Using Deep Denoising Sparse Autoencoder
    Bakhti, Yassine
    Fezza, Sid Ahmed
    Hamidouche, Wassim
    Deforges, Olivier
    IEEE ACCESS, 2019, 7 : 160397 - 160407
  • [38] DeepMTD: Moving Target Defense for Deep Visual Sensing against Adversarial Examples
    Song, Qun
    Yan, Zhenyu
    Tan, Rui
    ACM TRANSACTIONS ON SENSOR NETWORKS, 2022, 18 (01)
  • [39] Adversarial Attacks Against Deep Learning-Based Network Intrusion Detection Systems and Defense Mechanisms
    Zhang, Chaoyun
    Costa-Perez, Xavier
    Patras, Paul
    IEEE-ACM TRANSACTIONS ON NETWORKING, 2022, 30 (03) : 1294 - 1311
  • [40] Dynamic and Diverse Transformations for Defending Against Adversarial Examples
    Chen, Yongkang
    Zhang, Ming
    Li, Jin
    Kuang, Xiaohui
    Zhang, Xuhong
    Zhang, Han
    2022 IEEE INTERNATIONAL CONFERENCE ON TRUST, SECURITY AND PRIVACY IN COMPUTING AND COMMUNICATIONS, TRUSTCOM, 2022, : 976 - 983