On the Explainability of Natural Language Processing Deep Models

被引:41
|
作者
El Zini, Julia [1 ]
Awad, Mariette [1 ]
机构
[1] Amer Univ Beirut, Dept Elect & Comp Engn, POB 11-0236, Beirut 11072020, Lebanon
关键词
ExAI; NLP; language models; transformers; neural machine translation; transparent embedding models; explaining decisions; NEURAL-NETWORKS; GAME;
D O I
10.1145/3529755
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Despite their success, deep networks are used as black-box models with outputs that are not easily explainable during the learning and the prediction phases. This lack of interpretability is significantly limiting the adoption of such models in domains where decisions are critical such as the medical and legal fields. Recently, researchers have been interested in developing methods that help explain individual decisions and decipher the hidden representations of machine learning models in general and deep networks specifically. While there has been a recent explosion of work on Explainable Artificial Intelligence (ExAI) on deep models that operate on imagery and tabular data, textual datasets present new challenges to the ExAI community. Such challenges can be attributed to the lack of input structure in textual data, the use of word embeddings that add to the opacity of the models and the difficulty of the visualization of the inner workings of deep models when they are trained on textual data. Lately, methods have been developed to address the aforementioned challenges and present satisfactory explanations on Natural Language Processing (NLP) models. However, such methods are yet to be studied in a comprehensive framework where common challenges are properly stated and rigorous evaluation practices and metrics are proposed. Motivated to democratize ExAI methods in the NLP field, we present in this work a survey that studies model-agnostic as well as model-specific explainability methods on NLP models. Such methods can either develop inherently interpretable NLP models or operate on pre-trained models in a post hoc manner. We make this distinction and we further decompose the methods into three categories according to what they explain: (1) word embeddings (input level), (2) inner workings of NLP models (processing level), and (3) models' decisions (output level). We also detail the different evaluation approaches interpretability methods in the NLP field. Finally, we present a case-study on the well-known neural machine translation in an appendix, and we propose promising future research directions for ExAl in the NLP field.
引用
收藏
页数:31
相关论文
共 50 条
  • [21] Analysis of news sentiments using natural language processing and deep learning
    Mattia Vicari
    Mauro Gaspari
    AI & SOCIETY, 2021, 36 : 931 - 937
  • [22] Literature Survey of statistical, deep and reinforcement learning in Natural Language Processing
    Kaushik, Pranav
    Sharma, Akanksha Rai
    2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTING, COMMUNICATION AND AUTOMATION (ICCCA), 2017, : 350 - 354
  • [23] An Analysis of Early Use of Deep Learning Terms in Natural Language Processing
    Basic, B. Dalbelo
    di Buono, M. P.
    2020 43RD INTERNATIONAL CONVENTION ON INFORMATION, COMMUNICATION AND ELECTRONIC TECHNOLOGY (MIPRO 2020), 2020, : 1125 - 1129
  • [24] The journey from natural language processing to large language models: key insights for radiologists
    Salvatore Claudio Fanni
    Lorenzo Tumminello
    Valentina Formica
    Francesca Pia Caputo
    Gayane Aghakhanyan
    Ilaria Ambrosini
    Roberto Francischello
    Lorenzo Faggioni
    Dania Cioni
    Emanuele Neri
    Journal of Medical Imaging and Interventional Radiology, 11 (1):
  • [25] Encrypted malicious traffic detection based on natural language processing and deep learning
    Zang, Xiaodong
    Wang, Tongliang
    Zhang, Xinchang
    Gong, Jian
    Gao, Peng
    Zhang, Guowei
    COMPUTER NETWORKS, 2024, 250
  • [26] Pretrained Quantum-Inspired Deep Neural Network for Natural Language Processing
    Shi, Jinjing
    Chen, Tian
    Lai, Wei
    Zhang, Shichao
    Li, Xuelong
    IEEE TRANSACTIONS ON CYBERNETICS, 2024, 54 (10) : 5973 - 5985
  • [27] Large Language Models are Not Models of Natural Language: They are Corpus Models
    Veres, Csaba
    IEEE ACCESS, 2022, 10 : 61970 - 61979
  • [28] Detection of Pneumothorax with Deep Learning Models: Learning From Radiologist Labels vs Natural Language Processing Model Generated Labels
    Hallinan, James Thomas Patrick Decourcy
    Feng, Mengling
    Ng, Dianwen
    Sia, Soon Yiew
    Tiong, Vincent Tze Yang
    Jagmohan, Pooja
    Makmur, Andrew
    Thian, Yee Liang
    ACADEMIC RADIOLOGY, 2022, 29 (09) : 1350 - 1358
  • [29] Attention in Natural Language Processing
    Galassi, Andrea
    Lippi, Marco
    Torroni, Paolo
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (10) : 4291 - 4308
  • [30] Natural Language Processing Basics
    Arivazhagan, Naveen
    Van Vleck, Tielman T.
    CLINICAL JOURNAL OF THE AMERICAN SOCIETY OF NEPHROLOGY, 2023, 18 (03): : 400 - 401