Towards a Robust Deep Neural Network Against Adversarial Texts: A Survey

被引:31
作者
Wang, Wenqi [1 ,2 ]
Wang, Run [1 ,2 ]
Wang, Lina [1 ,2 ]
Wang, Zhibo [2 ,3 ]
Ye, Aoshuang [1 ,2 ]
机构
[1] Wuhan Univ, Key Lab Aerosp Informat Secur & Trusted Comp, Minist Educ, Wuhan 430072, Hubei, Peoples R China
[2] Wuhan Univ, Sch Cyber Sci & Engn, Wuhan 430072, Hubei, Peoples R China
[3] Zhejiang Univ, Sch Cyber Sci & Technol, Hangzhou 310027, Zhejiang, Peoples R China
基金
中国国家自然科学基金;
关键词
Task analysis; Perturbation methods; Natural language processing; Robustness; Information integrity; Analytical models; Sentiment analysis; Adversarial attacks and defenses; adversarial texts; robustness; deep neural networks; natural language processing;
D O I
10.1109/TKDE.2021.3117608
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks (DNNs) have achieved remarkable success in various tasks (e.g., image classification, speech recognition, and natural language processing (NLP)). However, researchers have demonstrated that DNN-based models are vulnerable to adversarial examples, which cause erroneous predictions by adding imperceptible perturbations into legitimate inputs. Recently, studies have revealed adversarial examples in the text domain, which could effectively evade various DNN-based text analyzers and further bring the threats of the proliferation of disinformation. In this paper, we give a comprehensive survey on the existing studies of adversarial techniques for generating adversarial texts written by both English and Chinese characters and the corresponding defense methods. More importantly, we hope that our work could inspire future studies to develop more robust DNN-based text analyzers against known and unknown adversarial techniques. We classify the existing adversarial techniques for crafting adversarial texts based on the perturbation units, helping to better understand the generation of adversarial texts and build robust models for defense. In presenting the taxonomy of adversarial attacks and defenses in the text domain, we introduce the adversarial techniques from the perspective of different NLP tasks. Finally, we discuss the existing challenges of adversarial attacks and defenses in texts and present the future research directions in this emerging and challenging field.
引用
收藏
页码:3159 / 3179
页数:21
相关论文
共 233 条
  • [1] Adi E., 2017, PROC INT C LEARN REP
  • [2] Adi Y., 2017, P INT C LEARN REPR
  • [3] Afchar D, 2018, IEEE INT WORKS INFOR
  • [4] Threat of Adversarial Attacks on Deep Learning in Computer Vision: A Survey
    Akhtar, Naveed
    Mian, Ajmal
    [J]. IEEE ACCESS, 2018, 6 : 14410 - 14430
  • [5] Alzantot M, 2018, 2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), P2890
  • [6] [Anonymous], 2004, arXiv
  • [7] [Anonymous], 2018, P 35 INT C MACH LEAR
  • [8] [Anonymous], 2016, ARXIV PREPRINT ARXIV
  • [9] [Anonymous], 2017, P 2 WORKSHOP EVALUAT
  • [10] [Anonymous], 2017, INT C LEARN REPR