Efficient Adversarial Training with Transferable Adversarial Examples

被引:78
作者
Zheng, Haizhong [1 ]
Zhang, Ziqi [1 ]
Gu, Juncheng [1 ]
Lee, Honglak [1 ]
Prakash, Atul [1 ]
机构
[1] Univ Michigan, Ann Arbor, MI 48109 USA
来源
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR) | 2020年
基金
美国国家科学基金会;
关键词
D O I
10.1109/CVPR42600.2020.00126
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Adversarial training is an effective defense method to protect classification models against adversarial attacks. However, one limitation of this approach is that it can require orders of magnitude additional training time due to high cost of generating strong adversarial examples during training. In this paper, we first show that there is high transferability between models from neighboring epochs in the same training process, i.e., adversarial examples from one epoch continue to be adversarial in subsequent epochs. Leveraging this property, we propose a novel method, Adversarial Training with Transferable Adversarial Examples (ATTA), that can enhance the robustness of trained models and greatly improve the training efficiency by accumulating adversarial perturbations through epochs. Compared to state-of-the-art adversarial training methods, AIM enhances adversarial accuracy by up to 7.2% on CIFAR10 and requires 12 similar to 14x less training time on MNIST and CIFAR10 datasets with comparable model robustness.
引用
收藏
页码:1178 / 1187
页数:10
相关论文
共 34 条
[1]  
[Anonymous], 2018, INT C MACH LEARN ICM
[2]  
[Anonymous], 2 INT C LEARN REPR I
[3]  
[Anonymous], 2017, INT C LEARN REPR ICL
[4]  
[Anonymous], 2016, BMVC
[5]  
[Anonymous], 2018, INT C LEARNING REPRE
[6]  
Baluja S, 2018, AAAI CONF ARTIF INTE, P2687
[7]  
Cai Qi-Zhi, 2018, INT JOINT C ART INT
[8]  
Carlini N., 2017, ACM WORKSH ART INT S, P3
[9]  
Carmon Yair, 2019, ADV NEURAL INFORM PR
[10]  
Cohen J, 2019, PR MACH LEARN RES, V97