Poison Ink: Robust and Invisible Backdoor Attack

被引:63
作者
Zhang, Jie [1 ]
Chen, Dongdong [2 ]
Huang, Qidong [1 ]
Liao, Jing [3 ]
Zhang, Weiming [1 ]
Feng, Huamin [4 ]
Hua, Gang [5 ]
Yu, Nenghai [1 ]
机构
[1] Univ Sci & Technol China, Sch Cyber Sci & Secur, Hefei 230026, Anhui, Peoples R China
[2] Microsoft Res, Redmond, WA 98052 USA
[3] City Univ Hong, Dept Comp Sci, Hong Kong, Peoples R China
[4] Beijing Elect Sci & Technol Inst, Beijing 100070, Peoples R China
[5] Wormpex AI Res LLC, Bellevue, WA 98004 USA
关键词
Toxicology; Ink; Training; Robustness; Data models; Training data; Task analysis; Backdoor attack; stealthiness; robustness; generality; flexibility;
D O I
10.1109/TIP.2022.3201472
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent research shows deep neural networks are vulnerable to different types of attacks, such as adversarial attacks, data poisoning attacks, and backdoor attacks. Among them, backdoor attacks are the most cunning and can occur in almost every stage of the deep learning pipeline. Backdoor attacks have attracted lots of interest from both academia and industry. However, most existing backdoor attack methods are visible or fragile to some effortless pre-processing such as common data transformations. To address these limitations, we propose a robust and invisible backdoor attack called "Poison Ink". Concretely, we first leverage the image structures as target poisoning areas and fill them with poison ink (information) to generate the trigger pattern. As the image structure can keep its semantic meaning during the data transformation, such a trigger pattern is inherently robust to data transformations. Then we leverage a deep injection network to embed such input-aware trigger pattern into the cover image to achieve stealthiness. Compared to existing popular backdoor attack methods, Poison Ink outperforms both in stealthiness and robustness. Through extensive experiments, we demonstrate that Poison Ink is not only general to different datasets and network architectures but also flexible for different attack scenarios. Besides, it also has very strong resistance against many state-of-the-art defense techniques.
引用
收藏
页码:5691 / 5705
页数:15
相关论文
共 69 条
[1]  
Baluja S, 2017, ADV NEUR IN, V30
[2]  
Barni M, 2019, IEEE IMAGE PROC, P101, DOI [10.1109/icip.2019.8802997, 10.1109/ICIP.2019.8802997]
[3]   STRONG DATA AUGMENTATION SANITIZES POISONING AND BACKDOOR ATTACKS WITHOUT AN ACCURACY TRADEOFF [J].
Borgnia, Eitan ;
Cherepanova, Valeriia ;
Fowl, Liam ;
Ghiasi, Amin ;
Geiping, Jonas ;
Goldblum, Micah ;
Goldstein, Tom ;
Gupta, Arjun .
2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, :3855-3859
[5]  
Carlini N, 2019, Arxiv, DOI arXiv:1902.06705
[6]   PCANet: A Simple Deep Learning Baseline for Image Classification? [J].
Chan, Tsung-Han ;
Jia, Kui ;
Gao, Shenghua ;
Lu, Jiwen ;
Zeng, Zinan ;
Ma, Yi .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2015, 24 (12) :5017-5032
[7]   Adversarial Attack Against Deep Saliency Models Powered by Non-Redundant Priors [J].
Che, Zhaohui ;
Borji, Ali ;
Zhai, Guangtao ;
Ling, Suiyi ;
Li, Jing ;
Tian, Yuan ;
Guo, Guodong ;
Le Callet, Patrick .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 :1973-1988
[8]  
Chen BY, 2018, Arxiv, DOI arXiv:1811.03728
[9]   Controllable Image Processing via Adaptive FilterBank Pyramid [J].
Chen, Dongdong ;
Fan, Qingnan ;
Liao, Jing ;
Aviles-Rivero, Angelica ;
Yuan, Lu ;
Yu, Nenghai ;
Hua, Gang .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 :8043-8054
[10]   FineFool: A novel DNN object contour attack on image recognition based on the attention perturbation adversarial technique [J].
Chen, Jinyin ;
Zheng, Haibin ;
Xiong, Hui ;
Chen, Ruoxi ;
Du, Tianyu ;
Hong, Zhen ;
Ji, Shouling .
COMPUTERS & SECURITY, 2021, 104