Label-Free Poisoning Attack Against Deep Unsupervised Domain Adaptation

被引:3
作者
Wang, Zhibo [1 ,2 ]
Liu, Wenxin [3 ]
Hu, Jiahui [1 ,2 ]
Guo, Hengchang [4 ]
Qin, Zhan [1 ,2 ]
Liu, Jian [1 ,2 ]
Ren, Kui [1 ,2 ]
机构
[1] Zhejiang Univ, Sch Cyber Sci & Technol, Hangzhou 310027, Zhejiang, Peoples R China
[2] ZJU Hangzhou Global Sci & Technol Innovat Ctr, Hangzhou 310027, Zhejiang, Peoples R China
[3] Ant Grp, Hangzhou 310058, Zhejiang, Peoples R China
[4] Wuhan Univ, Sch Cyber Sci & Engn, Wuhan 430072, Hubei, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Adaptation models; Data models; Training; Toxicology; Computational modeling; Training data; Deep learning; Poisoning attack; robustness; unsupervised domain adaptation;
D O I
10.1109/TDSC.2023.3286608
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Deep unsupervised domain adaptation (UDA) has significantly boosted the performance of deep models on different domains by transferring knowledge from a source domain to a target domain. However, its robustness against adversarial attacks has not been explored due to the challenges of highly non-convex deep models and different data distribution. In this article, we give the first attempt to analyze the vulnerability of deep UDA and propose a label-free poisoning attack (LFPA), which injects poisoning data into the training data to mislead adaptation between the two domains without ground truth in target domain. Specifically, we design an unsupervised adversarial loss as the attack goal, in which the pseudo-labels are used to approximate the ground-truth. Since retraining the model will gradually degrade the attack performance, we also add a regularization term to the unsupervised loss, which eliminates negative interactions between the training goal and the attack goal. To accelerate the craft of poisons, we select influential samples as the initial poisons and propose a fast reverse-mode optimization method which updates poisons according to the approximate truncated gradients. Experimental results on multiple state-of-the-art deep UDA methods demonstrate the effectiveness of the proposed LFPA and the high sensitivity of UDA to poisoning attacks.
引用
收藏
页码:1572 / 1586
页数:15
相关论文
共 40 条
[1]  
Alfeld S, 2016, AAAI CONF ARTIF INTE, P1452
[2]  
Biggio Battista., 2012, arXiv
[3]   Domain Adaptation Problems: A DASVM Classification Technique and a Circular Validation Strategy [J].
Bruzzone, Lorenzo ;
Marconcini, Mattia .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2010, 32 (05) :770-787
[4]   Towards Evaluating the Robustness of Neural Networks [J].
Carlini, Nicholas ;
Wagner, David .
2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2017, :39-57
[5]   Selective Transfer Machine for Personalized Facial Action Unit Detection [J].
Chu, Wen-Sheng ;
De la Torre, Fernando ;
Cohn, Jeffery F. .
2013 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2013, :3515-3522
[6]  
Collobert R, 2011, J MACH LEARN RES, V12, P2493
[7]  
Domke J., 2012, P 15 INT C ARTIFICIA
[8]  
Feng Q., 2019, P INT C NEUR INF PRO
[9]  
Franceschi L, 2017, PR MACH LEARN RES, V70
[10]  
Geiping J, 2021, Arxiv, DOI arXiv:2009.02276