Adversarial Robustness Across Representation Spaces

被引:6
作者
Awasthi, Pranjal [1 ]
Yu, George [1 ]
Ferng, Chun-Sung [1 ]
Tomkins, Andrew [1 ]
Juan, Da-Cheng [1 ]
机构
[1] Google Res, Mountain View, CA 94043 USA
来源
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021 | 2021年
关键词
D O I
10.1109/CVPR46437.2021.00752
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Adversarial robustness corresponds to the susceptibility of deep neural networks to imperceptible perturbations made at test time. In the context of image tasks, many algorithms have been proposed to make neural networks robust to adversarial perturbations made to the input pixels. These perturbations are typically measured in an 4, norm. However, robustness often holds only for the specific attack used for training. In this work we extend the above setting to consider the problem of training of deep neural networks that can be made simultaneously robust to perturbations applied in multiple natural representations spaces. For the case of image data, examples include the standard pixel representation as well as the representation in the discrete cosine transform (DCT) basis. We design a theoretically sound algorithm with formal guarantees for the above problem. Furthermore, our guarantees also hold when the goal is to require robustness with respect to multiple 4, norm based attacks. We then derive an efficient practical implementation and demonstrate the effectiveness of our approach on standard datasets for image classification.(1)
引用
收藏
页码:7604 / 7612
页数:9
相关论文
共 29 条
[1]  
Agarwal A., 2018, REDUCTIONS APPROACH, V80, P60
[2]   Defense against Universal Adversarial Perturbations [J].
Akhtar, Naveed ;
Liu, Jian ;
Mian, Ajmal .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :3389-3398
[3]  
[Anonymous], 2019, ARXIV190202918
[4]  
[Anonymous], 2018, Advances in Neural Information Processing Systems
[5]  
[Anonymous], 2017, COMMUN ACM, DOI DOI 10.1145/3065386
[6]  
Athalye A, 2018, PR MACH LEARN RES, V80
[7]  
Awasthi P., 2020, ADV NEURAL INFORM PR, V33
[8]   Towards Evaluating the Robustness of Neural Networks [J].
Carlini, Nicholas ;
Wagner, David .
2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2017, :39-57
[9]  
Cortes Corinna, 2020, Advances in Neural Information Processing Systems, V33
[10]  
COTTER A., 2019, ALGORITHMIC LEARNING, V98, P300