A Survey of Adversarial Defenses and Robustness in NLP

被引:36
|
作者
Goyal, Shreya [1 ]
Doddapaneni, Sumanth [1 ]
Khapra, Mitesh M. [1 ]
Ravindran, Balaraman [1 ]
机构
[1] Indian Inst Technol Madras, Bhupat & Jyoti Mehta Sch Biosci, Robert Bosch Ctr Data Sci & AI, Chennai 600036, Tamil Nadu, India
关键词
Adversarial attacks; adversarial defenses; perturbations; NLP; DEEP NEURAL-NETWORKS; COMPUTER VISION; ATTACKS;
D O I
10.1145/3593042
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
In the past few years, it has become increasingly evident that deep neural networks are not resilient enough to withstand adversarial perturbations in input data, leaving them vulnerable to attack. Various authors have proposed strong adversarial attacks for computer vision and Natural Language Processing (NLP) tasks. As a response, many defense mechanisms have also been proposed to prevent these networks from failing. The significance of defending neural networks against adversarial attacks lies in ensuring that the model's predictions remain unchanged even if the input data is perturbed. Several methods for adversarial defense in NLP have been proposed, catering to different NLP tasks such as text classification, named entity recognition, and natural language inference. Some of these methods not only defend neural networks against adversarial attacks but also act as a regularization mechanism during training, saving the model from overfitting. This survey aims to review the various methods proposed for adversarial defenses in NLP over the past few years by introducing a novel taxonomy. The survey also highlights the fragility of advanced deep neural networks in NLP and the challenges involved in defending them.
引用
收藏
页数:39
相关论文
共 50 条
  • [1] A Survey on Adversarial Attacks and Defenses for Deep Reinforcement Learning
    Liu A.-S.
    Guo J.
    Li S.-M.
    Xiao Y.-S.
    Liu X.-L.
    Tao D.-C.
    Jisuanji Xuebao/Chinese Journal of Computers, 2023, 46 (08): : 1553 - 1576
  • [2] Advances in Adversarial Attacks and Defenses in Computer Vision: A Survey
    Akhtar, Naveed
    Mian, Ajmal
    Kardan, Navid
    Shah, Mubarak
    IEEE ACCESS, 2021, 9 : 155161 - 155196
  • [3] A Survey on Efficient Methods for Adversarial Robustness
    Muhammad, Awais
    Bae, Sung-Ho
    IEEE ACCESS, 2022, 10 : 118815 - 118830
  • [4] How Deep Learning Sees the World: A Survey on Adversarial Attacks & Defenses
    Costa, Joana C.
    Roxo, Tiago
    Proenca, Hugo
    Inacio, Pedro Ricardo Morais
    IEEE ACCESS, 2024, 12 : 61113 - 61136
  • [5] A survey on adversarial attacks and defenses for object detection and their applications in autonomous vehicles
    Amirkhani, Abdollah
    Karimi, Mohammad Parsa
    Banitalebi-Dehkordi, Amin
    VISUAL COMPUTER, 2023, 39 (11) : 5293 - 5307
  • [6] A decade of adversarial examples: a survey on the nature and understanding of neural network non-robustness
    Trusov, A. V.
    Limonova, E. E.
    Arlazarov, V. V.
    COMPUTER OPTICS, 2025, 49 (02) : 222 - 252
  • [7] It Is All about Data: A Survey on the Effects of Data on Adversarial Robustness
    Xiong, Peiyu
    Tegegn, Michael
    Sarin, Jaskeerat Singh
    Pal, Shubhraneel
    Rubin, Julia
    ACM COMPUTING SURVEYS, 2024, 56 (07)
  • [8] Adversarial attacks and defenses in Speaker Recognition Systems: A survey
    Lan, Jiahe
    Zhang, Rui
    Yan, Zheng
    Wang, Jie
    Chen, Yu
    Hou, Ronghui
    JOURNAL OF SYSTEMS ARCHITECTURE, 2022, 127
  • [9] Generation and Countermeasures of adversarial examples on vision: a survey
    Liu, Jiangfan
    Li, Yishan
    Guo, Yanming
    Liu, Yu
    Tang, Jun
    Nie, Ying
    ARTIFICIAL INTELLIGENCE REVIEW, 2024, 57 (08)
  • [10] Adversarial attacks and defenses on text-to-image diffusion models: A survey
    Zhang, Chenyu
    Hu, Mingwang
    Li, Wenhui
    Wang, Lanjun
    INFORMATION FUSION, 2025, 114