Domain adaptation for textual adversarial defense via prompt-tuning

被引:0
作者
Li, Zhenglong [1 ]
Zhu, Yi [1 ,2 ,3 ]
Hua, Chenqi [1 ]
Li, Yun [1 ]
Yuan, Yunhao [1 ]
Qiang, Jipeng [1 ]
机构
[1] Yangzhou Univ, Sch Informat Engn, Yangzhou, Peoples R China
[2] Hefei Univ Technol, Key Lab Knowledge Engn Big Data, Minist Educ, Hefei, Peoples R China
[3] Hefei Univ Technol, Sch Comp Sci & Informat Engn, Hefei, Peoples R China
关键词
Textual adversarial defense; Adversarial attack; Domain adaptation; Prompt-tuning; Pre-trained Language Models; Verbalizer;
D O I
10.1016/j.neucom.2024.129192
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Most existing systems based on deep learning are vulnerable to adversarial attacks, i.e., strategically modified samples. These samples are generated by some imperceptible perturbations on different levels (sentence/word/character), which can fool deep models to give erroneous predictions. Recently, fine-tuning Pre-trained Language Models (PLMs) methods have achieved tremendous success in a series of downstream Natural Language Processing (NLP) tasks. However, recent studies indicate these methods need extensive adversarial examples for fine-tuning. In this paper, we propose a novel domain adaptation method for textual adversarial defense via prompt-tuning. Our method simulates original and adversarial text using source and target domains, then employs a prompt-tuning model to minimize domain deviations, treating textual adversarial defense as a cross-domain text classification task. To avoid noise and high time complexity from introducing external knowledge, we extract knowledge from the target domain and use hierarchical clustering and optimization strategies to fine-tune expansion words for the verbalizer in prompt-tuning. Experimental results show that the obvious improvement is obtained compared with other SOTA PLMs methods on textual adversarial defense.
引用
收藏
页数:10
相关论文
共 63 条
[1]  
Blitzer John., 2006, Proceedings of the 2006 Conference on Empirical Methods in Natural Language Processing, P120, DOI [10.3115/1610075.1610094, DOI 10.3115/1610075.1610094]
[2]  
Blohm Matthias., 2018, arXiv
[3]  
Branch HJ, 2022, Arxiv, DOI arXiv:2209.02128
[4]  
Brown TB, 2020, ADV NEUR IN, V33
[5]   KnowPrompt: Knowledge-aware Prompt-tuning with Synergistic Optimization for Relation Extraction [J].
Chen, Xiang ;
Zhang, Ningyu ;
Xie, Xin ;
Deng, Shumin ;
Yao, Yunzhi ;
Tan, Chuanqi ;
Huang, Fei ;
Si, Luo ;
Chen, Huajun .
PROCEEDINGS OF THE ACM WEB CONFERENCE 2022 (WWW'22), 2022, :2778-2788
[6]  
Devlin J, 2019, Arxiv, DOI arXiv:1810.04805
[7]  
Ding N., 2021, arXiv, DOI DOI 10.48550/ARXIV.2108.10604
[8]  
Ding N, 2021, Arxiv, DOI [arXiv:2111.01998, 10.48550/arXiv.2111.01998]
[9]  
Ebrahimi J, 2018, Arxiv, DOI arXiv:1712.06751
[10]  
Formento B, 2023, 17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, P1