A Survey of Adversarial Defenses and Robustness in NLP

被引:36
|
作者
Goyal, Shreya [1 ]
Doddapaneni, Sumanth [1 ]
Khapra, Mitesh M. [1 ]
Ravindran, Balaraman [1 ]
机构
[1] Indian Inst Technol Madras, Bhupat & Jyoti Mehta Sch Biosci, Robert Bosch Ctr Data Sci & AI, Chennai 600036, Tamil Nadu, India
关键词
Adversarial attacks; adversarial defenses; perturbations; NLP; DEEP NEURAL-NETWORKS; COMPUTER VISION; ATTACKS;
D O I
10.1145/3593042
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
In the past few years, it has become increasingly evident that deep neural networks are not resilient enough to withstand adversarial perturbations in input data, leaving them vulnerable to attack. Various authors have proposed strong adversarial attacks for computer vision and Natural Language Processing (NLP) tasks. As a response, many defense mechanisms have also been proposed to prevent these networks from failing. The significance of defending neural networks against adversarial attacks lies in ensuring that the model's predictions remain unchanged even if the input data is perturbed. Several methods for adversarial defense in NLP have been proposed, catering to different NLP tasks such as text classification, named entity recognition, and natural language inference. Some of these methods not only defend neural networks against adversarial attacks but also act as a regularization mechanism during training, saving the model from overfitting. This survey aims to review the various methods proposed for adversarial defenses in NLP over the past few years by introducing a novel taxonomy. The survey also highlights the fragility of advanced deep neural networks in NLP and the challenges involved in defending them.
引用
收藏
页数:39
相关论文
共 50 条
  • [31] AdvQuNN: A Methodology for Analyzing the Adversarial Robustness of Quanvolutional Neural Networks
    El Maouaki, Walid
    Marchisio, Alberto
    Said, Taoufik
    Bennai, Mohamed
    Shafique, Muhammad
    2024 IEEE INTERNATIONAL CONFERENCE ON QUANTUM SOFTWARE, IEEE QSW 2024, 2024, : 175 - 181
  • [32] Safety-critical computer vision: an empirical survey of adversarial evasion attacks and defenses on computer vision systems
    Charles Meyers
    Tommy Löfstedt
    Erik Elmroth
    Artificial Intelligence Review, 2023, 56 : 217 - 251
  • [33] Safety-critical computer vision: an empirical survey of adversarial evasion attacks and defenses on computer vision systems
    Meyers, Charles
    Lofstedt, Tommy
    Elmroth, Erik
    ARTIFICIAL INTELLIGENCE REVIEW, 2023, 56 (SUPPL 1) : 217 - 251
  • [34] Adversarial Attacks and Defenses in Images, Graphs and Text: A Review
    Xu, Han
    Ma, Yao
    Liu, Hao-Chen
    Deb, Debayan
    Liu, Hui
    Tang, Ji-Liang
    Jain, Anil K.
    INTERNATIONAL JOURNAL OF AUTOMATION AND COMPUTING, 2020, 17 (02) : 151 - 178
  • [35] Enhancing robustness of backdoor attacks against backdoor defenses
    Hu, Bin
    Guo, Kehua
    Ren, Sheng
    Fang, Hui
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 269
  • [36] Adversarial attacks in computer vision: a survey
    Li, Chao
    Wang, Handing
    Yao, Wen
    Jiang, Tingsong
    JOURNAL OF MEMBRANE COMPUTING, 2024, 6 (2) : 130 - 147
  • [37] Research Progress and Challenges on Application-Driven Adversarial Examples: A Survey
    Jiang, Wei
    He, Zhiyuan
    Zhan, Jinyu
    Pan, Weijia
    Adhikari, Deepak
    ACM TRANSACTIONS ON CYBER-PHYSICAL SYSTEMS, 2021, 5 (04)
  • [38] Fake News Detection via NLP is Vulnerable to Adversarial Attacks
    Zhou, Zhixuan
    Guan, Huankang
    Bhat, Meghana Moorthy
    Hsu, Justin
    PROCEEDINGS OF THE 11TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE (ICAART), VOL 2, 2019, : 794 - 800
  • [39] The Impact of Model Variations on the Robustness of Deep Learning Models in Adversarial Settings
    Juraev, Firuz
    Abuhamad, Mohammed
    Woo, Simon S.
    Thiruvathukal, George K.
    Abuhmed, Tamer
    2024 SILICON VALLEY CYBERSECURITY CONFERENCE, SVCC 2024, 2024,
  • [40] Adversarial Robustness for Deep Learning-Based Wildfire Prediction Models
    Ide, Ryo
    Yang, Lei
    FIRE-SWITZERLAND, 2025, 8 (02):