On the Explainability of Natural Language Processing Deep Models

被引:41
|
作者
El Zini, Julia [1 ]
Awad, Mariette [1 ]
机构
[1] Amer Univ Beirut, Dept Elect & Comp Engn, POB 11-0236, Beirut 11072020, Lebanon
关键词
ExAI; NLP; language models; transformers; neural machine translation; transparent embedding models; explaining decisions; NEURAL-NETWORKS; GAME;
D O I
10.1145/3529755
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Despite their success, deep networks are used as black-box models with outputs that are not easily explainable during the learning and the prediction phases. This lack of interpretability is significantly limiting the adoption of such models in domains where decisions are critical such as the medical and legal fields. Recently, researchers have been interested in developing methods that help explain individual decisions and decipher the hidden representations of machine learning models in general and deep networks specifically. While there has been a recent explosion of work on Explainable Artificial Intelligence (ExAI) on deep models that operate on imagery and tabular data, textual datasets present new challenges to the ExAI community. Such challenges can be attributed to the lack of input structure in textual data, the use of word embeddings that add to the opacity of the models and the difficulty of the visualization of the inner workings of deep models when they are trained on textual data. Lately, methods have been developed to address the aforementioned challenges and present satisfactory explanations on Natural Language Processing (NLP) models. However, such methods are yet to be studied in a comprehensive framework where common challenges are properly stated and rigorous evaluation practices and metrics are proposed. Motivated to democratize ExAI methods in the NLP field, we present in this work a survey that studies model-agnostic as well as model-specific explainability methods on NLP models. Such methods can either develop inherently interpretable NLP models or operate on pre-trained models in a post hoc manner. We make this distinction and we further decompose the methods into three categories according to what they explain: (1) word embeddings (input level), (2) inner workings of NLP models (processing level), and (3) models' decisions (output level). We also detail the different evaluation approaches interpretability methods in the NLP field. Finally, we present a case-study on the well-known neural machine translation in an appendix, and we propose promising future research directions for ExAl in the NLP field.
引用
收藏
页数:31
相关论文
共 50 条
  • [31] Natural Language Processing in Nephrology
    Vleck, Tielman T. Van
    Farrell, Douglas
    Chan, Lili
    ADVANCES IN CHRONIC KIDNEY DISEASE, 2022, 29 (05) : 465 - 471
  • [32] Natural Language Processing for Dialects of a Language: A Survey
    Joshi, Aditya
    Dabre, Raj
    Kanojia, Diptesh
    Li, Zhuang
    Zhan, Haolan
    Haffari, Gholamreza
    Dippold, Doris
    ACM COMPUTING SURVEYS, 2025, 57 (06)
  • [33] Thesauruses for natural language processing
    Kilgarriff, A
    2003 INTERNATIONAL CONFERENCE ON NATURAL LANGUAGE PROCESSING AND KNOWLEDGE ENGINEERING, PROCEEDINGS, 2003, : 5 - 13
  • [34] Recent Advances in Natural Language Processing via Large Pre-trained Language Models: A Survey
    Min, Bonan
    Ross, Hayley
    Sulem, Elior
    Ben Veyseh, Amir Pouran
    Nguyen, Thien Huu
    Sainz, Oscar
    Agirre, Eneko
    Heintz, Ilana
    Roth, Dan
    ACM COMPUTING SURVEYS, 2024, 56 (02)
  • [35] Safe Pretraining of Deep Language Models in a Synthetic Pseudo-Language
    Gorbacheva, T. E.
    Bondarenko, I. Y.
    DOKLADY MATHEMATICS, 2023, 108 (SUPPL 2) : S494 - S502
  • [36] Applications of Natural Language Processing and Large Language Models for Social Determinants of Health: Protocol for a Systematic Review
    Rajwal, Swati
    Zhang, Ziyuan
    Chen, Yankai
    Rogers, Hannah
    Sarker, Abeed
    Xiao, Yunyu
    JMIR RESEARCH PROTOCOLS, 2025, 14
  • [37] Accelerated evidence synthesis in orthopaedics—the roles of natural language processing, expert annotation and large language models
    Bálint Zsidai
    Janina Kaarre
    Ann-Sophie Hilkert
    Eric Narup
    Eric Hamrin Senorski
    Alberto Grassi
    Olufemi R. Ayeni
    Volker Musahl
    Christophe Ley
    Elmar Herbst
    Michael T. Hirschmann
    Sebastian Kopf
    Romain Seil
    Thomas Tischer
    Kristian Samuelsson
    Robert Feldt
    Journal of Experimental Orthopaedics, 10
  • [38] Assessment of Deep Natural Language Processing in Ascertaining Oncologic Outcomes From Radiology Reports
    Kehl, Kenneth L.
    Elmarakeby, Haitham
    Nishino, Mizuki
    Van Allen, Eliezer M.
    Lepisto, Eva M.
    Hassett, Michael J.
    Johnson, Bruce E.
    Schrag, Deborah
    JAMA ONCOLOGY, 2019, 5 (10) : 1421 - 1429
  • [39] An approach to detect offence in Memes using Natural Language Processing(NLP) and Deep learning
    Giri, Roushan Kumar
    Gupta, Subhash Chandra
    Gupta, Umesh Kumar
    2021 INTERNATIONAL CONFERENCE ON COMPUTER COMMUNICATION AND INFORMATICS (ICCCI), 2021,
  • [40] A Phishing-Attack-Detection Model Using Natural Language Processing and Deep Learning
    Benavides-Astudillo, Eduardo
    Fuertes, Walter
    Sanchez-Gordon, Sandra
    Nunez-Agurto, Daniel
    Rodriguez-Galan, German
    APPLIED SCIENCES-BASEL, 2023, 13 (09):