Indirect Invisible Poisoning Attacks on Domain Adaptation

被引:12
作者
Wu, Jun [1 ]
He, Jingrui [1 ]
机构
[1] Univ Illinois, Champaign, IL 61820 USA
来源
KDD '21: PROCEEDINGS OF THE 27TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING | 2021年
基金
美国食品与农业研究所; 美国国家科学基金会;
关键词
Domain Adaptation; Domain Discrepancy; Poisoning Attack;
D O I
10.1145/3447548.3467214
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Unsupervised domain adaptation has been successfully applied across multiple high-impact applications, since it improves the generalization performance of a learning algorithm when the source and target domains are related. However, the adversarial vulnerability of domain adaptation models has largely been neglected. Most existing unsupervised domain adaptation algorithms might be easily fooled by an adversary, resulting in deteriorated prediction performance on the target domain, when transferring the knowledge from a maliciously manipulated source domain. To demonstrate the adversarial vulnerability of existing domain adaptation techniques, in this paper, we propose a generic data poisoning attack framework named I2Attack for domain adaptation with the following properties: (1) perceptibly unnoticeable: all the poisoned inputs are natural-looking; (2) adversarially indirect: only source examples are maliciously manipulated; (3) algorithmically invisible: both source classification error and marginal domain discrepancy between source and target domains will not increase. Specifically, it aims to degrade the overall prediction performance on the target domain by maximizing the label-informed domain discrepancy over both input feature space and class-label space between source and target domains. Within this framework, a family of practical poisoning attacks are presented to fool the existing domain adaptation algorithms associated with different discrepancy measures. Extensive experiments on various domain adaptation benchmarks confirm the effectiveness and computational efficiency of our proposed I2Attack framework.
引用
收藏
页码:1852 / 1862
页数:11
相关论文
共 40 条
  • [1] No Place to Hide: Catching Fraudulent Entities in Tensors
    Ban, Yikun
    Liu, Xin
    Duan, Yitao
    Liu, Xue
    Xu, Wei
    [J]. WEB CONFERENCE 2019: PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE (WWW 2019), 2019, : 83 - 93
  • [2] A theory of learning from different domains
    Ben-David, Shai
    Blitzer, John
    Crammer, Koby
    Kulesza, Alex
    Pereira, Fernando
    Vaughan, Jennifer Wortman
    [J]. MACHINE LEARNING, 2010, 79 (1-2) : 151 - 175
  • [3] BIGGIO B., 2012, arXiv preprint arXiv:1206.6389, P1467
  • [4] Towards Evaluating the Robustness of Neural Networks
    Carlini, Nicholas
    Wagner, David
    [J]. 2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2017, : 39 - 57
  • [5] CHEN XY, 2019, PR MACH LEARN RES, V97
  • [6] Fallah A, 2020, PR MACH LEARN RES, V108, P1082
  • [7] Finn C, 2017, PR MACH LEARN RES, V70
  • [8] Ganin Y, 2015, PR MACH LEARN RES, V37, P1180
  • [9] Ge R., 2015, PMLR, V40, P797
  • [10] Scatter Component Analysis: A Unified Framework for Domain Adaptation and Domain Generalization
    Ghifary, Muhammad
    Balduzzi, David
    Kleijn, W. Bastiaan
    Zhang, Mengjie
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (07) : 1414 - 1430