BadNets: Evaluating Backdooring Attacks on Deep Neural Networks

被引:658
作者
Gu, Tianyu [1 ]
Liu, Kang [1 ]
Dolan-Gavitt, Brendan [2 ]
Garg, Siddharth [1 ]
机构
[1] NYU, Dept Elect & Comp Engn, New York, NY 11002 USA
[2] NYU, Dept Comp Sci & Engn, New York, NY 11002 USA
基金
美国国家科学基金会;
关键词
Computer security; machine learning; neural networks;
D O I
10.1109/ACCESS.2019.2909068
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep learning-based techniques have achieved state-of-the-art performance on a wide variety of recognition and classification tasks. However, these networks are typically computationally expensive to train, requiring weeks of computation on many GPUs; as a result, many users outsource the training procedure to the cloud or rely on pre-trained models that are then fine-tuned for a specific task. In this paper, we show that the outsourced training introduces new security risks: an adversary can create a maliciously trained network (a backdoored neural network, or a BadNet) that has the state-of-the-art performance on the user's training and validation samples but behaves badly on specific attacker-chosen inputs. We first explore the properties of BadNets in a toy example, by creating a backdoored handwritten digit classifier. Next, we demonstrate backdoors in a more realistic scenario by creating a U.S. street sign classifier that identifies stop signs as speed limits when a special sticker is added to the stop sign; we then show in addition that the backdoor in our U.S. street sign detector can persist even if the network is later retrained for another task and cause a drop in an accuracy of 25% on average when the backdoor trigger is present. These results demonstrate that backdoors in neural networks are both powerful and-because the behavior of neural networks is difficult to explicate-stealthy. This paper provides motivation for further research into techniques for verifying and inspecting neural networks, just as we have developed tools for verifying and debugging software.
引用
收藏
页码:47230 / 47244
页数:15
相关论文
共 54 条
[11]  
[Anonymous], ADV NEURAL INFORM PR, DOI DOI 10.1109/TPAMI.2016.2577031
[12]  
[Anonymous], CONVEXIFIED CONVOLUT
[13]  
[Anonymous], 2014, P INT C LEARNING REP
[14]  
[Anonymous], 2015, P ICLR
[15]  
[Anonymous], 1988, NIPS
[16]  
[Anonymous], 2014, P ADV NEUR INF PROC
[17]  
[Anonymous], GOOGL CLOUD MACH LEA
[18]  
[Anonymous], PROC 17TH INT IEEE C
[19]  
Bahdanau D., 2015, P INT C LEARN REPR, P1
[20]  
Barreno M, 2006, P 2006 ACM S INF COM, P16