GradDiv: Adversarial Robustness of Randomized Neural Networks via Gradient Diversity Regularization

被引:20
作者
Lee, Sungyoon [1 ]
Kim, Hoki [2 ]
Lee, Jaewook [2 ]
机构
[1] Korea Inst Adv Study KIAS, Ctr Artificial Intelligence & Nat Sci, Seoul 02455, South Korea
[2] Seoul Natl Univ, Dept Ind Engn, Seoul 08826, South Korea
基金
新加坡国家研究基金会;
关键词
Adversarial robustness; defense against adversarial attacks; randomized neural networks; directional analysis;
D O I
10.1109/TPAMI.2022.3169217
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep learning is vulnerable to adversarial examples. Many defenses based on randomized neural networks have been proposed to solve the problem, but fail to achieve robustness against attacks using proxy gradients such as the Expectation over Transformation (EOT) attack. We investigate the effect of the adversarial attacks using proxy gradients on randomized neural networks and demonstrate that it highly relies on the directional distribution of the loss gradients of the randomized neural network. We show in particular that proxy gradients are less effective when the gradients are more scattered. To this end, we propose Gradient Diversity (GradDiv) regularizations that minimize the concentration of the gradients to build a robust randomized neural network. Our experiments on MNIST, CIFAR10, and STL10 show that our proposed GradDiv regularizations improve the adversarial robustness of randomized neural networks against a variety of state-of-the-art attack methods. Moreover, our method efficiently reduces the transferability among sample models of randomized neural networks.
引用
收藏
页码:2645 / 2651
页数:7
相关论文
共 34 条
[1]  
Andriushchenko Maksym, 2020, Computer Vision - ECCV 2020. 16th European Conference. Proceedings. Lecture Notes in Computer Science (LNCS 12368), P484, DOI 10.1007/978-3-030-58592-1_29
[2]  
[Anonymous], 2014, P INT C LEARNING REP
[3]  
Athalye A, 2018, PR MACH LEARN RES, V80
[4]  
Athalye A, 2018, PR MACH LEARN RES, V80
[5]  
Banerjee A, 2005, J MACH LEARN RES, V6, P1345
[6]   Eynard-Mehta theorem, schur process, and their pfaffian analogs [J].
Borodin, A ;
Rains, EM .
JOURNAL OF STATISTICAL PHYSICS, 2005, 121 (3-4) :291-317
[7]  
Brendel W., 2019, ADV NEUR IN, V32
[8]   Towards Evaluating the Robustness of Neural Networks [J].
Carlini, Nicholas ;
Wagner, David .
2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2017, :39-57
[9]  
Coates A., 2011, AISTATS, P215
[10]  
Croce F, 2019, 25TH AMERICAS CONFERENCE ON INFORMATION SYSTEMS (AMCIS 2019)