A Survey of Adversarial Defenses and Robustness in NLP

被引:36
|
作者
Goyal, Shreya [1 ]
Doddapaneni, Sumanth [1 ]
Khapra, Mitesh M. [1 ]
Ravindran, Balaraman [1 ]
机构
[1] Indian Inst Technol Madras, Bhupat & Jyoti Mehta Sch Biosci, Robert Bosch Ctr Data Sci & AI, Chennai 600036, Tamil Nadu, India
关键词
Adversarial attacks; adversarial defenses; perturbations; NLP; DEEP NEURAL-NETWORKS; COMPUTER VISION; ATTACKS;
D O I
10.1145/3593042
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
In the past few years, it has become increasingly evident that deep neural networks are not resilient enough to withstand adversarial perturbations in input data, leaving them vulnerable to attack. Various authors have proposed strong adversarial attacks for computer vision and Natural Language Processing (NLP) tasks. As a response, many defense mechanisms have also been proposed to prevent these networks from failing. The significance of defending neural networks against adversarial attacks lies in ensuring that the model's predictions remain unchanged even if the input data is perturbed. Several methods for adversarial defense in NLP have been proposed, catering to different NLP tasks such as text classification, named entity recognition, and natural language inference. Some of these methods not only defend neural networks against adversarial attacks but also act as a regularization mechanism during training, saving the model from overfitting. This survey aims to review the various methods proposed for adversarial defenses in NLP over the past few years by introducing a novel taxonomy. The survey also highlights the fragility of advanced deep neural networks in NLP and the challenges involved in defending them.
引用
收藏
页数:39
相关论文
共 50 条
  • [21] Understanding deep learning defenses against adversarial examples through visualizations for dynamic risk assessment
    Echeberria-Barrio, Xabier
    Gil-Lerchundi, Amaia
    Egana-Zubia, Jon
    Orduna-Urrutia, Raul
    NEURAL COMPUTING & APPLICATIONS, 2022, 34 (23) : 20477 - 20490
  • [22] Adversarial Attacks and Defenses in Machine Learning-Empowered Communication Systems and Networks: A Contemporary Survey
    Wang, Yulong
    Sun, Tong
    Li, Shenghong
    Yuan, Xin
    Ni, Wei
    Hossain, Ekram
    Vincent Poor, H.
    IEEE COMMUNICATIONS SURVEYS AND TUTORIALS, 2023, 25 (04): : 2245 - 2298
  • [23] On Saliency Maps and Adversarial Robustness
    Mangla, Puneet
    Singh, Vedant
    Balasubramanian, Vineeth N.
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2020, PT II, 2021, 12458 : 272 - 288
  • [24] Adversarial attacks and defenses in person search: A systematic mapping study and taxonomy
    Andrade, Eduardo de O.
    Guerin, Joris
    Viterbo, Jose
    Sampaio, Igor Garcia Ballhausen
    IMAGE AND VISION COMPUTING, 2024, 148
  • [25] Rethinking maximum-margin softmax for adversarial robustness
    Hassanin, Mohammed
    Moustafa, Nour
    Tahtali, Murat
    Choo, Kim-Kwang Raymond
    COMPUTERS & SECURITY, 2022, 116
  • [26] Adversarial Robustness of Model Sets
    Megyeri, Istvan
    Hegedus, Istvan
    Jelasity, Mark
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [27] Evaluating the Adversarial Robustness of Text Classifiers in Hyperdimensional Computing
    Moraliyage, Harsha
    Kahawala, Sachin
    De Silva, Daswin
    Alahakoon, Damminda
    2022 15TH INTERNATIONAL CONFERENCE ON HUMAN SYSTEM INTERACTION (HSI), 2022,
  • [28] Privacy and Robustness in Federated Learning: Attacks and Defenses
    Lyu, Lingjuan
    Yu, Han
    Ma, Xingjun
    Chen, Chen
    Sun, Lichao
    Zhao, Jun
    Yang, Qiang
    Yu, Philip S.
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (07) : 8726 - 8746
  • [29] A Survey of Attacks and Defenses for Deep Neural Networks
    Machooka, Daniel
    Yuan, Xiaohong
    Esterline, Albert
    2023 IEEE INTERNATIONAL CONFERENCE ON CYBER SECURITY AND RESILIENCE, CSR, 2023, : 254 - 261
  • [30] Adversarial attacks and defenses for large language models (LLMs): methods, frameworks & challenges
    Kumar, Pranjal
    INTERNATIONAL JOURNAL OF MULTIMEDIA INFORMATION RETRIEVAL, 2024, 13 (03)