Trojan Attack on Deep Generative Models in Autonomous Driving

被引:16
作者
Ding, Shaohua [1 ]
Tian, Yulong [1 ]
Xu, Fengyuan [1 ]
Li, Qun [2 ]
Zhong, Sheng [1 ]
机构
[1] Nanjing Univ, State Key Lab Novel Software Technol, Nanjing, Peoples R China
[2] Coll William & Mary, Williamsburg, VA USA
来源
SECURITY AND PRIVACY IN COMMUNICATION NETWORKS, SECURECOMM, PT I | 2019年 / 304卷
关键词
Deep generative models; Trojan attacks; Autonomous driving; Data poisoning;
D O I
10.1007/978-3-030-37228-6_15
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep generative models (DGMs) have empowered unprecedented innovations in many application domains. However, their security has not been thoroughly assessed when deploying such models in practice, especially in those mission-critical tasks like autonomous driving. In this work, we draw attention to a new attack surface of DGMs, which is the data used in the training phase. We demonstrate that the training data poisoning, the injection of specially-crafted data, are able to teach Trojan behaviors to a DGM without influencing the original training goal. Such Trojan attack will be activated after model deployment only if certain rare triggers are present in an input. For example, a rain-removal DGM after poisoning can, while removing raindrops in input images, change a traffic light from red to green if this traffic light has a specific appearance (i.e. a trigger). Clearly severe consequences can occur if such poisoned model is deployed on vehicle. Our study shows that launching our Trojan attack is feasible on different DGM categories designed for the autonomous driving scenario, and existing defense methods cannot effectively defeat it. We also introduce a concealing technique to make our data poisoning more inconspicuous during the training. In the end, we propose some potential defense strategies inspiring future explorations.
引用
收藏
页码:299 / 318
页数:20
相关论文
共 32 条
[1]  
[Anonymous], 2017, ARXIV171205526
[2]  
[Anonymous], Bloomberg News
[3]  
Brock A, 2018, ARXIV
[4]  
Chen Bryant, 2018, ARXIV181103728
[5]   Residual-Guide Network for Single Image Deraining [J].
Fan, Zhiwen ;
Wu, Huafeng ;
Fu, Xueyang ;
Huang, Yue ;
Ding, Xinghao .
PROCEEDINGS OF THE 2018 ACM MULTIMEDIA CONFERENCE (MM'18), 2018, :1751-1759
[6]   Generative Adversarial Networks [J].
Goodfellow, Ian ;
Pouget-Abadie, Jean ;
Mirza, Mehdi ;
Xu, Bing ;
Warde-Farley, David ;
Ozair, Sherjil ;
Courville, Aaron ;
Bengio, Yoshua .
COMMUNICATIONS OF THE ACM, 2020, 63 (11) :139-144
[7]  
Gu T., 2017, P MACH LEARN COMP SE
[8]  
Hayes Jamie, 2019, Proceedings on Privacy Enhancing Technologies, V2019, P133, DOI 10.2478/popets-2019-0008
[9]   Image-to-Image Translation with Conditional Adversarial Networks [J].
Isola, Phillip ;
Zhu, Jun-Yan ;
Zhou, Tinghui ;
Efros, Alexei A. .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :5967-5976
[10]   Model-Reuse Attacks on Deep Learning Systems [J].
Ji, Yujie ;
Zhang, Xinyang ;
Ji, Shouling ;
Luo, Xiapu ;
Wang, Ting .
PROCEEDINGS OF THE 2018 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY (CCS'18), 2018, :349-363