Towards a Robust Deep Neural Network Against Adversarial Texts: A Survey

被引:31
作者
Wang, Wenqi [1 ,2 ]
Wang, Run [1 ,2 ]
Wang, Lina [1 ,2 ]
Wang, Zhibo [2 ,3 ]
Ye, Aoshuang [1 ,2 ]
机构
[1] Wuhan Univ, Key Lab Aerosp Informat Secur & Trusted Comp, Minist Educ, Wuhan 430072, Hubei, Peoples R China
[2] Wuhan Univ, Sch Cyber Sci & Engn, Wuhan 430072, Hubei, Peoples R China
[3] Zhejiang Univ, Sch Cyber Sci & Technol, Hangzhou 310027, Zhejiang, Peoples R China
基金
中国国家自然科学基金;
关键词
Task analysis; Perturbation methods; Natural language processing; Robustness; Information integrity; Analytical models; Sentiment analysis; Adversarial attacks and defenses; adversarial texts; robustness; deep neural networks; natural language processing;
D O I
10.1109/TKDE.2021.3117608
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks (DNNs) have achieved remarkable success in various tasks (e.g., image classification, speech recognition, and natural language processing (NLP)). However, researchers have demonstrated that DNN-based models are vulnerable to adversarial examples, which cause erroneous predictions by adding imperceptible perturbations into legitimate inputs. Recently, studies have revealed adversarial examples in the text domain, which could effectively evade various DNN-based text analyzers and further bring the threats of the proliferation of disinformation. In this paper, we give a comprehensive survey on the existing studies of adversarial techniques for generating adversarial texts written by both English and Chinese characters and the corresponding defense methods. More importantly, we hope that our work could inspire future studies to develop more robust DNN-based text analyzers against known and unknown adversarial techniques. We classify the existing adversarial techniques for crafting adversarial texts based on the perturbation units, helping to better understand the generation of adversarial texts and build robust models for defense. In presenting the taxonomy of adversarial attacks and defenses in the text domain, we introduce the adversarial techniques from the perspective of different NLP tasks. Finally, we discuss the existing challenges of adversarial attacks and defenses in texts and present the future research directions in this emerging and challenging field.
引用
收藏
页码:3159 / 3179
页数:21
相关论文
共 233 条
  • [61] Gan WC, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), P6065
  • [62] Adversarial Perturbations Fool Deepfake Detectors
    Gandhi, Apurva
    Jain, Shomik
    [J]. 2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [63] Black-box Generation of Adversarial Text Sequences to Evade Deep Learning Classifiers
    Gao, Ji
    Lanchantin, Jack
    Soffa, Mary Lou
    Qi, Yanjun
    [J]. 2018 IEEE SYMPOSIUM ON SECURITY AND PRIVACY WORKSHOPS (SPW 2018), 2018, : 50 - 56
  • [64] Garg S, 2020, PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), P6174
  • [65] Gehring J, 2017, PR MACH LEARN RES, V70
  • [66] Gil Y, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P1373
  • [67] Gilmer J., 2018, arXiv:1807.06732
  • [68] Gilmer J, 2018, Arxiv, DOI arXiv:1807.06732
  • [69] Glockner M, 2018, PROCEEDINGS OF THE 56TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 2, P650
  • [70] Gong H., 2018, PROC INT C LEARN REP