Invisible Poison: A Blackbox Clean Label Backdoor Attack to Deep Neural Networks

被引:31
作者
Ning, Rui [1 ]
Li, Jiang [2 ]
Xin, Chunsheng [1 ,2 ]
Wu, Hongyi [1 ,2 ]
机构
[1] Old Dominion Univ, Sch Cybersecur, Norfolk, VA 23529 USA
[2] Old Dominion Univ, Dept Elect & Comp Engn, Norfolk, VA 23529 USA
来源
IEEE CONFERENCE ON COMPUTER COMMUNICATIONS (IEEE INFOCOM 2021) | 2021年
基金
美国国家科学基金会;
关键词
Deep Learning; Neural Backdoor; Security;
D O I
10.1109/INFOCOM42981.2021.9488902
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
This paper reports a new clean-label data poisoning backdoor attack, named Invisible Poison, which stealthily and aggressively plants a backdoor in neural networks. It converts a regular trigger to a noised trigger that can be easily concealed inside images for training NN, with the objective to plant a backdoor that can be later activated by the trigger. Compared with existing data poisoning backdoor attacks, this newfound attack has the following distinct properties. First, it is a blackbox attack, requiring zero-knowledge of the target model. Second, this attack utilizes "invisible poison" to achieve stealthiness where the trigger is disguised as `noise', and thus can easily evade human inspection. On the other hand, this noised trigger remains effective in the feature space to poison training data. Third, the attack is practical and aggressive. A backdoor can be effectively planted with a small amount of poisoned data and is robust to most data augmentation methods during training. The attack is fully tested on multiple benchmark datasets including MNIST, Cifar10, and ImageNet10, as well as application specific data sets such as Yahoo Adblocker and GTSRB. Two countermeasures, namely Supervised and Unsupervised Poison Sample Detection, are introduced to defend the attack.
引用
收藏
页数:10
相关论文
共 35 条
[1]  
Al-Qizwini M, 2017, IEEE INT VEH SYM, P89, DOI 10.1109/IVS.2017.7995703
[2]  
Amodei D, 2016, PR MACH LEARN RES, V48
[3]  
Arjovsky M., 2017, ARXIV170107875
[4]  
Barni M, 2019, IEEE IMAGE PROC, P101, DOI [10.1109/ICIP.2019.8802997, 10.1109/icip.2019.8802997]
[5]  
Boski M, 2017, 2017 10TH INTERNATIONAL WORKSHOP ON MULTIDIMENSIONAL (ND) SYSTEMS (NDS)
[6]  
Chen Xinyun, 2017, ARXIV171205526
[7]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[8]  
Din ZA, 2020, PROCEEDINGS OF THE 2020 USENIX ANNUAL TECHNICAL CONFERENCE, P387
[9]   Backdoor Embedding in Convolutional Neural Network Models via Invisible Perturbation [J].
Zhong, Haoti ;
Liao, Cong ;
Squicciarini, Anna Cinzia ;
Zhu, Sencun ;
Miller, David .
PROCEEDINGS OF THE TENTH ACM CONFERENCE ON DATA AND APPLICATION SECURITY AND PRIVACY, CODASPY 2020, 2020, :97-108
[10]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778