A Survey of Adversarial Defenses and Robustness in NLP

被引:36
|
作者
Goyal, Shreya [1 ]
Doddapaneni, Sumanth [1 ]
Khapra, Mitesh M. [1 ]
Ravindran, Balaraman [1 ]
机构
[1] Indian Inst Technol Madras, Bhupat & Jyoti Mehta Sch Biosci, Robert Bosch Ctr Data Sci & AI, Chennai 600036, Tamil Nadu, India
关键词
Adversarial attacks; adversarial defenses; perturbations; NLP; DEEP NEURAL-NETWORKS; COMPUTER VISION; ATTACKS;
D O I
10.1145/3593042
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
In the past few years, it has become increasingly evident that deep neural networks are not resilient enough to withstand adversarial perturbations in input data, leaving them vulnerable to attack. Various authors have proposed strong adversarial attacks for computer vision and Natural Language Processing (NLP) tasks. As a response, many defense mechanisms have also been proposed to prevent these networks from failing. The significance of defending neural networks against adversarial attacks lies in ensuring that the model's predictions remain unchanged even if the input data is perturbed. Several methods for adversarial defense in NLP have been proposed, catering to different NLP tasks such as text classification, named entity recognition, and natural language inference. Some of these methods not only defend neural networks against adversarial attacks but also act as a regularization mechanism during training, saving the model from overfitting. This survey aims to review the various methods proposed for adversarial defenses in NLP over the past few years by introducing a novel taxonomy. The survey also highlights the fragility of advanced deep neural networks in NLP and the challenges involved in defending them.
引用
收藏
页数:39
相关论文
共 50 条
  • [41] Does Simple Trump Complex? Comparing Strategies for Adversarial Robustness in DNNs
    Brooks, William
    Davel, Marelie H.
    Mouton, Coenraad
    ARTIFICIAL INTELLIGENCE RESEARCH, SACAIR 2024, 2025, 2326 : 253 - 269
  • [42] Robustness Against Adversarial Attacks in Neural Networks Using Incremental Dissipativity
    Aquino, Bernardo
    Rahnama, Arash
    Seiler, Peter
    Lin, Lizhen
    Gupta, Vijay
    IEEE CONTROL SYSTEMS LETTERS, 2022, 6 : 2341 - 2346
  • [43] Error Resiliency and Adversarial Robustness in Convolutional Neural Networks: An Empirical Analysis
    Barbareschi, Mario
    Barone, Salvatore
    Casola, Valentina
    Della Torca, Salvatore
    INTERNET OF THINGS, IFIPIOT 2024, 2025, 737 : 149 - 160
  • [44] Adversarial Deep Learning: A Survey on Adversarial Attacks and Defense Mechanisms on Image Classification
    Khamaiseh, Samer Y.
    Bagagem, Derek
    Al-Alaj, Abdullah
    Mancino, Mathew
    Alomari, Hakam W.
    IEEE ACCESS, 2022, 10 : 102266 - 102291
  • [45] On the robustness of skeleton detection against adversarial attacks
    Bai, Xiuxiu
    Yang, Ming
    Liu, Zhe
    NEURAL NETWORKS, 2020, 132 : 416 - 427
  • [46] Bolstering Adversarial Robustness with Latent Disparity Regularization
    Schwartz, David
    Ditzler, Gregory
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [47] Adversarial Robustness of Phishing Email Detection Models
    Gholampour, Parisa Mehdi
    Verma, Rakesh M.
    PROCEEDINGS OF THE 9TH ACM INTERNATIONAL WORKSHOP ON SECURITY AND PRIVACY ANALYTICS, IWSPA 2023, 2023, : 67 - 76
  • [48] Adversarial Robustness of Sparse Local Lipschitz Predictors
    Muthukumar, Ramchandran
    Sulam, Jeremias
    SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 2023, 5 (04): : 920 - 948
  • [49] On the Relationship between Generalization and Robustness to Adversarial Examples
    Pedraza, Anibal
    Deniz, Oscar
    Bueno, Gloria
    SYMMETRY-BASEL, 2021, 13 (05):
  • [50] Robustness Against Adversarial Attacks Using Dimensionality
    Chattopadhyay, Nandish
    Chatterjee, Subhrojyoti
    Chattopadhyay, Anupam
    SECURITY, PRIVACY, AND APPLIED CRYPTOGRAPHY ENGINEERING, SPACE 2021, 2022, 13162 : 226 - 241