On the Explainability of Natural Language Processing Deep Models

被引:41
|
作者
El Zini, Julia [1 ]
Awad, Mariette [1 ]
机构
[1] Amer Univ Beirut, Dept Elect & Comp Engn, POB 11-0236, Beirut 11072020, Lebanon
关键词
ExAI; NLP; language models; transformers; neural machine translation; transparent embedding models; explaining decisions; NEURAL-NETWORKS; GAME;
D O I
10.1145/3529755
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Despite their success, deep networks are used as black-box models with outputs that are not easily explainable during the learning and the prediction phases. This lack of interpretability is significantly limiting the adoption of such models in domains where decisions are critical such as the medical and legal fields. Recently, researchers have been interested in developing methods that help explain individual decisions and decipher the hidden representations of machine learning models in general and deep networks specifically. While there has been a recent explosion of work on Explainable Artificial Intelligence (ExAI) on deep models that operate on imagery and tabular data, textual datasets present new challenges to the ExAI community. Such challenges can be attributed to the lack of input structure in textual data, the use of word embeddings that add to the opacity of the models and the difficulty of the visualization of the inner workings of deep models when they are trained on textual data. Lately, methods have been developed to address the aforementioned challenges and present satisfactory explanations on Natural Language Processing (NLP) models. However, such methods are yet to be studied in a comprehensive framework where common challenges are properly stated and rigorous evaluation practices and metrics are proposed. Motivated to democratize ExAI methods in the NLP field, we present in this work a survey that studies model-agnostic as well as model-specific explainability methods on NLP models. Such methods can either develop inherently interpretable NLP models or operate on pre-trained models in a post hoc manner. We make this distinction and we further decompose the methods into three categories according to what they explain: (1) word embeddings (input level), (2) inner workings of NLP models (processing level), and (3) models' decisions (output level). We also detail the different evaluation approaches interpretability methods in the NLP field. Finally, we present a case-study on the well-known neural machine translation in an appendix, and we propose promising future research directions for ExAl in the NLP field.
引用
收藏
页数:31
相关论文
共 50 条
  • [1] Fast compression and optimization of deep learning models for natural language processing
    Pietron, Marcin
    Karwatowski, Michal
    Wielgosz, Maciej
    Duda, Jerzy
    2019 SEVENTH INTERNATIONAL SYMPOSIUM ON COMPUTING AND NETWORKING WORKSHOPS (CANDARW 2019), 2019, : 162 - 168
  • [2] An introduction to Deep Learning in Natural Language Processing: Models, techniques, and tools
    Lauriola, Ivano
    Lavelli, Alberto
    Aiolli, Fabio
    NEUROCOMPUTING, 2022, 470 : 443 - 456
  • [3] Context-Sensitive Visualization of Deep Learning Natural Language Processing Models
    Dunn, Andrew
    Inkpen, Diana
    Andonie, Razvan
    2021 25TH INTERNATIONAL CONFERENCE INFORMATION VISUALISATION (IV): AI & VISUAL ANALYTICS & DATA SCIENCE, 2021, : 170 - 175
  • [4] A Survey on Adversarial Text Attacks on Deep Learning Models in Natural Language Processing
    Deepan, S.
    Torres-Cruz, Fred
    Placido-Lerma, Ruben L.
    Udhayakumar, R.
    Anuradha, S.
    Kapila, Dhiraj
    PROCEEDINGS OF THE 5TH INTERNATIONAL CONFERENCE ON DATA SCIENCE, MACHINE LEARNING AND APPLICATIONS, VOL 1, ICDSMLA 2023, 2025, 1273 : 1059 - 1067
  • [5] Deep Learning Methods in Natural Language Processing
    Flores, Alexis Stalin Alulema
    APPLIED TECHNOLOGIES (ICAT 2019), PT II, 2020, 1194 : 92 - 107
  • [6] Deep Learning Techniques for Natural Language Processing
    Rodzin, Sergey
    Bova, Victoria
    Kravchenko, Yury
    Rodzina, Lada
    ARTIFICIAL INTELLIGENCE TRENDS IN SYSTEMS, VOL 2, 2022, 502 : 121 - 130
  • [7] Natural Language Processing: An Overview of Models, Transformers and Applied Practices
    Canchila, Santiago
    Meneses-Eraso, Carlos
    Casanoves-Boix, Javier
    Cortes-Pellicer, Pascual
    Castello-Sirvent, Fernando
    COMPUTER SCIENCE AND INFORMATION SYSTEMS, 2024, 21 (03) : 1097 - 1145
  • [8] Overview of Character-Based Models for Natural Language Processing
    Adel, Heike
    Asgari, Ehsaneddin
    Schuetze, Hinrich
    COMPUTATIONAL LINGUISTICS AND INTELLIGENT TEXT PROCESSING (CICLING 2017), PT I, 2018, 10761 : 3 - 16
  • [9] Processing Natural Language on Embedded Devices: How Well Do Transformer Models Perform?
    Sarkar, Souvika
    Babar, Mohammad Fakhruddin
    Hassan, Md Mahadi
    Hasan, Monowar
    Santu, Shubhra Kanti Karmaker
    PROCEEDINGS OF THE 15TH ACM/SPEC INTERNATIONAL CONFERENCE ON PERFORMANCE ENGINEERING, ICPE 2024, 2024, : 211 - 222
  • [10] Fairness Certification for Natural Language Processing and Large Language Models
    Freiberger, Vincent
    Buchmann, Erik
    INTELLIGENT SYSTEMS AND APPLICATIONS, VOL 1, INTELLISYS 2024, 2024, 1065 : 606 - 624