Data filtering for efficient adversarial training

被引:5
作者
Chen, Erh-Chung [1 ]
Lee, Che-Rung [1 ]
机构
[1] Natl Tsing Hua Univ, Dept Comp Sci, Hsinchu, Taiwan
关键词
Adversarial training; Data pruning; Multiple objective optimization;
D O I
10.1016/j.patcog.2024.110394
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Adversarial training has been considered to be one of the most effective strategies to defend against adversarial attacks. Most existing adversarial training methods have shown a trade-off between training cost and robustness. This paper explores a new optimization direction from training data to reduce the computational cost of adversarial training without scarifying robustness. First, we show that some adversarial examples are less important, meaning that removing them does not hurt the robustness. Second, we propose a method to identify insignificant adversarial examples at a minimal cost. Third, we demonstrate that our approach can be integrated with other adversarial training frameworks with few modifications. The experimental results show that combined with previous works, our approach not only reduces about 20% of computational cost on the CIFAR10 and CIFAR100 datasets but also improves about 1.5% natural accuracy. With less computational cost, it achieves 58.22%, 30.68%, and 41.92% robust accuracy on CIFAR10, CIFAR100, and ImageNet datasets respectively, which are higher than those of the original methods.
引用
收藏
页数:12
相关论文
共 66 条
[1]  
Addepalli S., 2021, ICML 2021 WORKSHOP A
[2]   Adversarial example detection for DNN models: a review and experimental comparison [J].
Aldahdooh, Ahmed ;
Hamidouche, Wassim ;
Fezza, Sid Ahmed ;
Deforges, Olivier .
ARTIFICIAL INTELLIGENCE REVIEW, 2022, 55 (06) :4403-4462
[3]  
Andriushchenko M, 2020, Arxiv, DOI arXiv:1912.00049
[4]  
Athalye A, 2018, PR MACH LEARN RES, V80
[5]  
Brown TB, 2018, Arxiv, DOI arXiv:1809.08352
[6]  
Bochkovskiy A, 2020, Arxiv, DOI arXiv:2004.10934
[7]  
Cai QZ, 2018, Arxiv, DOI arXiv:1805.04807
[8]   Towards Evaluating the Robustness of Neural Networks [J].
Carlini, Nicholas ;
Wagner, David .
2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2017, :39-57
[9]   Audio Adversarial Examples: Targeted Attacks on Speech-to-Text [J].
Carlini, Nicholas ;
Wagner, David .
2018 IEEE SYMPOSIUM ON SECURITY AND PRIVACY WORKSHOPS (SPW 2018), 2018, :1-7
[10]  
Carmon Y, 2019, ADV NEUR IN, V32