Feature Distillation in Deep Attention Network Against Adversarial Examples

被引:7
|
作者
Chen, Xin [1 ,2 ,3 ,4 ]
Weng, Jian [5 ]
Deng, Xiaoling [1 ,2 ,3 ,4 ]
Luo, Weiqi [5 ]
Lan, Yubin [1 ,2 ,3 ,4 ]
Tian, Qi [6 ]
机构
[1] South China Agr Univ, Coll Elect Engn, Coll Artificial Intelligence, Guangzhou 510642, Peoples R China
[2] Natl Ctr Int Collaborat Res Precis Agr Aviat Pest, Guangzhou 510642, Peoples R China
[3] Guangdong Lab Lingnan Modern Agr, Guangzhou 510642, Peoples R China
[4] Guangdong Engn Technol Res Ctr Smart Agr, Guangzhou 510642, Peoples R China
[5] Jinan Univ, Coll Informat Sci & Technol, Guangzhou 510632, Peoples R China
[6] Huawei Cloud & AI, Shenzhen 518100, Peoples R China
基金
中国国家自然科学基金;
关键词
Perturbation methods; Robustness; Image coding; Training; Neural networks; Task analysis; Research and development; Adversarial examples; deep attention mechanism; frequency decomposition; image classification;
D O I
10.1109/TNNLS.2021.3113342
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks (DNNs) are easily fooled by adversarial examples. Most existing defense strategies defend against adversarial examples based on full information of whole images. In reality, one possible reason as to why humans are not sensitive to adversarial perturbations is that the human visual mechanism often concentrates on most important regions of images. A deep attention mechanism has been applied in many computer fields and has achieved great success. Attention modules are composed of an attention branch and a trunk branch. The encoder/decoder architecture in the attention branch has potential of compressing adversarial perturbations. In this article, we theoretically prove that attention modules can compress adversarial perturbations by destroying potential linear characteristics of DNNs. Considering the distribution characteristics of adversarial perturbations in different frequency bands, we design and compare three types of attention modules based on frequency decomposition and reorganization to defend against adversarial examples. Moreover, we find that our designed attention modules can obtain high classification accuracies on clean images by locating attention regions more accurately. Experimental results on the CIFAR and ImageNet dataset demonstrate that frequency reorganization in attention modules can not only achieve good robustness to adversarial perturbations, but also obtain comparable, even higher classification, accuracies on clean images. Moreover, our proposed attention modules can be integrated with existing defense strategies as components to further improve adversarial robustness.
引用
收藏
页码:3691 / 3705
页数:15
相关论文
共 50 条
  • [41] The Problem of the Adversarial Examples in Deep Learning
    Zhang S.-S.
    Zuo X.
    Liu J.-W.
    Jisuanji Xuebao/Chinese Journal of Computers, 2019, 42 (08): : 1886 - 1904
  • [42] A Synergetic Attack against Neural Network Classifiers combining Backdoor and Adversarial Examples
    Liu, Guanxiong
    Khalil, Issa
    Khreishah, Abdallah
    Phan, NhatHai
    2021 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2021, : 834 - 846
  • [43] Self-Attention Context Network: Addressing the Threat of Adversarial Attacks for Hyperspectral Image Classification
    Xu, Yonghao
    Du, Bo
    Zhang, Liangpei
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 : 8671 - 8685
  • [44] Adversarial Attack Against Deep Saliency Models Powered by Non-Redundant Priors
    Che, Zhaohui
    Borji, Ali
    Zhai, Guangtao
    Ling, Suiyi
    Li, Jing
    Tian, Yuan
    Guo, Guodong
    Le Callet, Patrick
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 : 1973 - 1988
  • [45] Generating Semantic Adversarial Examples via Feature Manipulation in Latent Space
    Wang, Shuo
    Chen, Shangyu
    Chen, Tianle
    Nepal, Surya
    Rudolph, Carsten
    Grobler, Marthie
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (12) : 17070 - 17084
  • [46] Rallying Adversarial Techniques against Deep Learning for Network Security
    Clements, Joseph
    Yang, Yuzhe
    Sharma, Ankur A.
    Hu, Hongxin
    Lao, Yingjie
    2021 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI 2021), 2021,
  • [47] Explaining Image Misclassification in Deep Learning via Adversarial Examples
    Haffar, Rami
    Jebreel, Najeeb Moharram
    Domingo-Ferrer, Josep
    Sanchez, David
    MODELING DECISIONS FOR ARTIFICIAL INTELLIGENCE (MDAI 2021), 2021, 12898 : 323 - 334
  • [48] Towards Transferable Adversarial Attack Against Deep Face Recognition
    Zhong, Yaoyao
    Deng, Weihong
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2021, 16 : 1452 - 1466
  • [49] Defending Against Deep Learning-Based Traffic Fingerprinting Attacks With Adversarial Examples
    Hayden, Blake
    Walsh, Timothy
    Barton, Armon
    ACM TRANSACTIONS ON PRIVACY AND SECURITY, 2025, 28 (01)
  • [50] Improving Deep Learning Model Robustness Against Adversarial Attack by Increasing the Network Capacity
    Marchetti, Marco
    Ho, Edmond S. L.
    ADVANCES IN CYBERSECURITY, CYBERCRIMES, AND SMART EMERGING TECHNOLOGIES, 2023, 4 : 85 - 96