Adversarial robustness via noise injection in smoothed models

被引:0
作者
Yaniv Nemcovsky
Evgenii Zheltonozhskii
Chaim Baskin
Brian Chmiel
Alex M. Bronstein
Avi Mendelson
机构
[1] Technion,Department of Computer Science
[2] An Intel Company,Habana Labs
[3] Technion,Department of Electrical Engineering
来源
Applied Intelligence | 2023年 / 53卷
关键词
Neural networks; Computer vision; Adversarial robustness; Adversarial examples; Noise injection; Randomized smoothing;
D O I
暂无
中图分类号
学科分类号
摘要
Deep neural networks are known to be vulnerable to malicious perturbations. Current methods for improving adversarial robustness make use of either implicit or explicit regularization, with the latter is usually based on adversarial training. Randomized smoothing, the averaging of the classifier outputs over a random distribution centered in the sample, has been shown to guarantee a classifier’s performance subject to bounded perturbations of the input. In this work, we study the application of randomized smoothing to improve performance on unperturbed data and increase robustness to adversarial attacks. We propose to combine smoothing along with adversarial training and randomization approaches, and find that doing so significantly improves the resilience compared to the baseline. We examine our method’s performance on common white-box (FGSM, PGD) and black-box (transferable attack and NAttack) attacks on CIFAR-10 and CIFAR-100, and determine that for a low number of iterations, smoothing provides a significant performance boost that persists even for perturbations with a high attack norm, 𝜖. For example, under a PGD-10 attack on CIFAR-10 using Wide-ResNet28-4, we achieve 60.3% accuracy for infinity norm 𝜖∞=8/255\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$\epsilon _{\infty }=\nicefrac {8}{255}$\end{document} and 13.1% accuracy for 𝜖∞=35/255\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$\epsilon _{\infty }=\nicefrac {35}{255}$\end{document} – outperforming previous art by 3% and 6%, respectively. We achieve nearly twice the accuracy on 𝜖∞=35/255\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$\epsilon _{\infty }=\nicefrac {35}{255}$\end{document} and even more so for perturbations with higher infinity norm. A https://github.com/yanemcovsky/SIAM of the proposed method is provided.
引用
收藏
页码:9483 / 9498
页数:15
相关论文
共 10 条
[1]  
Liu A(2021)Training robust deep neural networks via adversarial noise propagation IEEE Trans Image Process 30 5769-5781
[2]  
Liu X(2020)Enresnet: Resnets ensemble via the feynman–kac formalism for adversarial defense and beyond SIAM J Math Data Sci 2 559-582
[3]  
Yu H(undefined)undefined undefined undefined undefined-undefined
[4]  
Zhang C(undefined)undefined undefined undefined undefined-undefined
[5]  
Liu Q(undefined)undefined undefined undefined undefined-undefined
[6]  
Tao D(undefined)undefined undefined undefined undefined-undefined
[7]  
Wang B(undefined)undefined undefined undefined undefined-undefined
[8]  
Yuan B(undefined)undefined undefined undefined undefined-undefined
[9]  
Shi Z(undefined)undefined undefined undefined undefined-undefined
[10]  
Osher SJ(undefined)undefined undefined undefined undefined-undefined