Gradient Shielding: Towards Understanding Vulnerability of Deep Neural Networks

被引:44
作者
Gu, Zhaoquan [1 ]
Hu, Weixiong [1 ]
Zhang, Chuanjing [1 ]
Lu, Hui [1 ]
Yin, Lihua [1 ]
Wang, Le [1 ]
机构
[1] Guangzhou Univ, Cyberspace Inst Adv Technol, Guangzhou 510006, Peoples R China
来源
IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING | 2021年 / 8卷 / 02期
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Gradient shielding; adversarial example; deep neural networks; vulnerability; ROBUSTNESS; ATTACKS;
D O I
10.1109/TNSE.2020.2996738
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
Deep neural networks (DNNs) have been widely adopted but they are vulnerable to intentionally crafted adversarial examples. Various attack methods against DNNs have been proposed, yet there still lacks theoretical explanation of adversarial examples. In this paper, we aim to understand adversarial examples from the attacking process and we assume adding perturbations to the key/sensitive regions of the image could fool image classification DNNs. We propose gradient shielding to verify the assumption which ignores insensitive information during generating adversarial examples. Specifically, we propose interactive gradient shielding (IGS) method which selects sensitive regions and then applies gradient-based attack. To remove region selection, we propose adaptive gradient shielding (AGS) method which ignores insensitive gradients automatically. We conduct extensive experiments to evaluate the performance and the results also corroborate our perspective. With this method, we won the first place in IJCAI-AAAC 2019 Non-targeted Adversarial Attack competition.
引用
收藏
页码:921 / 932
页数:12
相关论文
共 65 条
[1]   Threat of Adversarial Attacks on Deep Learning in Computer Vision: A Survey [J].
Akhtar, Naveed ;
Mian, Ajmal .
IEEE ACCESS, 2018, 6 :14410-14430
[2]  
[Anonymous], 2017, P INT C LEARN REPR
[3]  
[Anonymous], 2015, ARXIV151105122
[4]  
[Anonymous], 2018, P INT C LEARN REPR
[5]  
[Anonymous], 2016, NAACL HLT 2016 2016, DOI DOI 10.18653/V1/N16-1062
[6]  
[Anonymous], P INT C MACH LEARN
[7]  
Athalye Anish, 2018, P MACHINE LEARNING R, V80
[8]   Towards Evaluating the Robustness of Neural Networks [J].
Carlini, Nicholas ;
Wagner, David .
2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2017, :39-57
[9]  
Chen PY, 2017, PROCEEDINGS OF THE 10TH ACM WORKSHOP ON ARTIFICIAL INTELLIGENCE AND SECURITY, AISEC 2017, P15, DOI 10.1145/3128572.3140448
[10]   ArcFace: Additive Angular Margin Loss for Deep Face Recognition [J].
Deng, Jiankang ;
Guo, Jia ;
Xue, Niannan ;
Zafeiriou, Stefanos .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :4685-4694