Parameterizing Activation Functions for Adversarial Robustness

被引:11
作者
Dai, Sihui [1 ]
Mahloujifar, Saeed [1 ]
Mittal, Prateek [1 ]
机构
[1] Princeton Univ, Elect & Comp Engn, Princeton, NJ 08544 USA
来源
2022 43RD IEEE SYMPOSIUM ON SECURITY AND PRIVACY WORKSHOPS (SPW 2022) | 2022年
基金
美国国家科学基金会;
关键词
D O I
10.1109/SPW54247.2022.9833884
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep neural networks are known to be vulnerable to adversarially perturbed inputs. A commonly used defense is adversarial training, whose performance is influenced by model architecture. While previous works have studied the impact of varying model width and depth on robustness, the impact of using learnable parametric activation functions (PAFs) has not been studied. We study how using learnable PAFs can improve robustness in conjunction with adversarial training. We first ask the question: Can changing activation function shape improve robustness? To address this, we choose a set of PAFs with parameters that allow us to independently control behavior on negative inputs, inputs near zero, and positive inputs. Using these PAFs, we train models using adversarial training with fixed PAF shape parameter values. We find that all regions of PAF shape influence the robustness of obtained models, however only variation in certain regions (inputs near zero, positive inputs) can improve robustness over ReLU. We then combine learnable PAFs with adversarial training and analyze robust performance. We find that choice of activation function can significantly impact the robustness of the trained model. We find that only certain PAFs, such as smooth PAFs, are able to improve robustness significantly over ReLU. Overall, our work puts into context the importance of activation functions in adversarially trained models.
引用
收藏
页码:80 / 87
页数:8
相关论文
共 36 条
[1]   Square Attack: A Query-Efficient Black-Box Adversarial Attack via Random Search [J].
Andriushchenko, Maksym ;
Croce, Francesco ;
Flammarion, Nicolas ;
Hein, Matthias .
COMPUTER VISION - ECCV 2020, PT XXIII, 2020, 12368 :484-501
[2]  
Brendel W., 2018, ICLR
[3]  
Bubeck Sebastien, 2021, ARXIV
[4]   Towards Evaluating the Robustness of Neural Networks [J].
Carlini, Nicholas ;
Wagner, David .
2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2017, :39-57
[5]  
Carmon Y, 2019, 33 C NEURAL INFORM P, V32
[6]  
Croce F., 2020, arXiv
[7]  
Croce F, 2020, PR MACH LEARN RES, V119
[8]  
Dugas C, 2001, ADV NEUR IN, V13, P472
[9]  
Glorot X., 2011, P 14 INT C ART INT S, P315, DOI DOI 10.1002/ECS2.1832
[10]  
Godfrey LB, 2019, IEEE SYS MAN CYBERN, P3006, DOI [10.1109/smc.2019.8913972, 10.1109/SMC.2019.8913972]