Feature Distillation in Deep Attention Network Against Adversarial Examples

被引:7
|
作者
Chen, Xin [1 ,2 ,3 ,4 ]
Weng, Jian [5 ]
Deng, Xiaoling [1 ,2 ,3 ,4 ]
Luo, Weiqi [5 ]
Lan, Yubin [1 ,2 ,3 ,4 ]
Tian, Qi [6 ]
机构
[1] South China Agr Univ, Coll Elect Engn, Coll Artificial Intelligence, Guangzhou 510642, Peoples R China
[2] Natl Ctr Int Collaborat Res Precis Agr Aviat Pest, Guangzhou 510642, Peoples R China
[3] Guangdong Lab Lingnan Modern Agr, Guangzhou 510642, Peoples R China
[4] Guangdong Engn Technol Res Ctr Smart Agr, Guangzhou 510642, Peoples R China
[5] Jinan Univ, Coll Informat Sci & Technol, Guangzhou 510632, Peoples R China
[6] Huawei Cloud & AI, Shenzhen 518100, Peoples R China
基金
中国国家自然科学基金;
关键词
Perturbation methods; Robustness; Image coding; Training; Neural networks; Task analysis; Research and development; Adversarial examples; deep attention mechanism; frequency decomposition; image classification;
D O I
10.1109/TNNLS.2021.3113342
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks (DNNs) are easily fooled by adversarial examples. Most existing defense strategies defend against adversarial examples based on full information of whole images. In reality, one possible reason as to why humans are not sensitive to adversarial perturbations is that the human visual mechanism often concentrates on most important regions of images. A deep attention mechanism has been applied in many computer fields and has achieved great success. Attention modules are composed of an attention branch and a trunk branch. The encoder/decoder architecture in the attention branch has potential of compressing adversarial perturbations. In this article, we theoretically prove that attention modules can compress adversarial perturbations by destroying potential linear characteristics of DNNs. Considering the distribution characteristics of adversarial perturbations in different frequency bands, we design and compare three types of attention modules based on frequency decomposition and reorganization to defend against adversarial examples. Moreover, we find that our designed attention modules can obtain high classification accuracies on clean images by locating attention regions more accurately. Experimental results on the CIFAR and ImageNet dataset demonstrate that frequency reorganization in attention modules can not only achieve good robustness to adversarial perturbations, but also obtain comparable, even higher classification, accuracies on clean images. Moreover, our proposed attention modules can be integrated with existing defense strategies as components to further improve adversarial robustness.
引用
收藏
页码:3691 / 3705
页数:15
相关论文
共 50 条
  • [21] Feature-Based Adversarial Training for Deep Learning Models Resistant to Transferable Adversarial Examples
    Ryu, Gwonsang
    Choi, Daeseon
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2022, E105D (05) : 1039 - 1049
  • [22] Defending Network IDS against Adversarial Examples with Continual Learning
    Kozal, Jedrzej
    Zwolinska, Justyna
    Klonowski, Marek
    Wozniak, Michal
    2023 23RD IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS, ICDMW 2023, 2023, : 60 - 69
  • [23] A Framework for Enhancing Deep Neural Networks Against Adversarial Malware
    Li, Deqiang
    Li, Qianmu
    Ye, Yanfang
    Xu, Shouhuai
    IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, 2021, 8 (01): : 736 - 750
  • [24] EEJE: Two-Step Input Transformation for Robust DNN Against Adversarial Examples
    Choi, Seok-Hwan
    Shin, Jinmyeong
    Liu, Peng
    Choi, Yoon-Ho
    IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, 2021, 8 (02): : 908 - 920
  • [25] ROBUSTNESS OF DEEP NEURAL NETWORKS IN ADVERSARIAL EXAMPLES
    Teng, Da
    Song, Xiao m
    Gong, Guanghong
    Han, Liang
    INTERNATIONAL JOURNAL OF INDUSTRIAL ENGINEERING-THEORY APPLICATIONS AND PRACTICE, 2017, 24 (02): : 123 - 133
  • [26] Conditional Generative Adversarial Network-Based Image Denoising for Defending Against Adversarial Attack
    Zhang, Haibo
    Sakurai, Kouichi
    IEEE ACCESS, 2021, 9 : 169031 - 169043
  • [27] Neuron Selecting: Defending Against Adversarial Examples in Deep Neural Networks
    Zhang, Ming
    Li, Hu
    Kuang, Xiaohui
    Pang, Ling
    Wu, Zhendong
    INFORMATION AND COMMUNICATIONS SECURITY (ICICS 2019), 2020, 11999 : 613 - 629
  • [28] MoNet: Impressionism As A Defense Against Adversarial Examples
    Ge, Huangyi
    Chau, Sze Yiu
    Li, Ninghui
    2020 SECOND IEEE INTERNATIONAL CONFERENCE ON TRUST, PRIVACY AND SECURITY IN INTELLIGENT SYSTEMS AND APPLICATIONS (TPS-ISA 2020), 2020, : 246 - 255
  • [29] Toward Invisible Adversarial Examples Against DNN-Based Privacy Leakage for Internet of Things
    Ding, Xuyang
    Zhang, Shuai
    Song, Mengkai
    Ding, Xiaocong
    Li, Fagen
    IEEE INTERNET OF THINGS JOURNAL, 2021, 8 (02): : 802 - 812
  • [30] Adversarial Examples Detection With Bayesian Neural Network
    Li, Yao
    Tang, Tongyi
    Hsieh, Cho-Jui
    Lee, Thomas C. M.
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2024, 8 (05): : 3654 - 3664