On the Explainability of Natural Language Processing Deep Models

被引:41
|
作者
El Zini, Julia [1 ]
Awad, Mariette [1 ]
机构
[1] Amer Univ Beirut, Dept Elect & Comp Engn, POB 11-0236, Beirut 11072020, Lebanon
关键词
ExAI; NLP; language models; transformers; neural machine translation; transparent embedding models; explaining decisions; NEURAL-NETWORKS; GAME;
D O I
10.1145/3529755
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Despite their success, deep networks are used as black-box models with outputs that are not easily explainable during the learning and the prediction phases. This lack of interpretability is significantly limiting the adoption of such models in domains where decisions are critical such as the medical and legal fields. Recently, researchers have been interested in developing methods that help explain individual decisions and decipher the hidden representations of machine learning models in general and deep networks specifically. While there has been a recent explosion of work on Explainable Artificial Intelligence (ExAI) on deep models that operate on imagery and tabular data, textual datasets present new challenges to the ExAI community. Such challenges can be attributed to the lack of input structure in textual data, the use of word embeddings that add to the opacity of the models and the difficulty of the visualization of the inner workings of deep models when they are trained on textual data. Lately, methods have been developed to address the aforementioned challenges and present satisfactory explanations on Natural Language Processing (NLP) models. However, such methods are yet to be studied in a comprehensive framework where common challenges are properly stated and rigorous evaluation practices and metrics are proposed. Motivated to democratize ExAI methods in the NLP field, we present in this work a survey that studies model-agnostic as well as model-specific explainability methods on NLP models. Such methods can either develop inherently interpretable NLP models or operate on pre-trained models in a post hoc manner. We make this distinction and we further decompose the methods into three categories according to what they explain: (1) word embeddings (input level), (2) inner workings of NLP models (processing level), and (3) models' decisions (output level). We also detail the different evaluation approaches interpretability methods in the NLP field. Finally, we present a case-study on the well-known neural machine translation in an appendix, and we propose promising future research directions for ExAl in the NLP field.
引用
收藏
页数:31
相关论文
共 50 条
  • [41] Accelerated evidence synthesis in orthopaedics-the roles of natural language processing, expert annotation and large language models
    Zsidai, Balint
    Kaarre, Janina
    Hilkert, Ann-Sophie
    Narup, Eric
    Senorski, Eric Hamrin
    Grassi, Alberto
    Ayeni, Olufemi R.
    Musahl, Volker
    Ley, Christophe
    Herbst, Elmar
    Hirschmann, Michael T.
    Kopf, Sebastian
    Seil, Romain
    Tischer, Thomas
    Samuelsson, Kristian
    Feldt, Robert
    JOURNAL OF EXPERIMENTAL ORTHOPAEDICS, 2023, 10 (01)
  • [42] Cyberbullying Detection: Hybrid Models Based on Machine Learning and Natural Language Processing Techniques
    Raj, Chahat
    Agarwal, Ayush
    Bharathy, Gnana
    Narayan, Bhuva
    Prasad, Mukesh
    ELECTRONICS, 2021, 10 (22)
  • [43] Techniques Comparison for Natural Language Processing
    Iosifova, Olena
    Iosifov, Ievgen
    Rolik, Oleksandr
    Sokolov, Volodymyr
    MOMLET+DS 2020: MODERN MACHINE LEARNING TECHNOLOGIES AND DATA SCIENCE WORKSHOP, 2020, 2631
  • [44] Natural language processing in medicine: A review
    Locke, Saskia
    Bashall, Anthony
    Al-Adely, Sarah
    Moore, John
    Wilson, Anthony
    Kitchen, Gareth B.
    TRENDS IN ANAESTHESIA AND CRITICAL CARE, 2021, 38 : 4 - 9
  • [45] Resources for Turkish natural language processing
    Coltekin, Cagri
    Dogruoz, A. Seza
    Cetinoglu, Ozlem
    LANGUAGE RESOURCES AND EVALUATION, 2023, 57 (01) : 449 - 488
  • [46] Natural Language Processing for Ancient Greek
    Stopponi, Silvia
    Pedrazzini, Nilo
    Peels-Matthey, Saskia
    McGillivray, Barbara
    Nissim, Malvina
    DIACHRONICA, 2024, 41 (03) : 414 - 435
  • [47] Reusable Toolkit for Natural Language Processing in an Ambient Intelligence Environment
    Saeedi, Sirwe
    Fong, A. C. M.
    Gupta, Ajay
    Carr, Steve
    2022 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (SSCI), 2022, : 429 - 435
  • [48] MULTI-CLASS AUTOMATED SPEECH LANGUAGE RECOGNITION USING NATURAL LANGUAGE PROCESSING WITH OPTIMAL DEEP LEARNING MODEL
    Al-anazi, Reema g.
    Alqahtani, Hamed
    Alzaidi, Muhammad swaileh a.
    Alanazi, Meshari h.
    AL Sultan, Hanan
    Alrowaily, Amal f.
    Aljabri, Jawhara
    Alqudah, Assal
    FRACTALS-COMPLEX GEOMETRY PATTERNS AND SCALING IN NATURE AND SOCIETY, 2025,
  • [49] An overview and empirical comparison of natural language processing (NLP) models and an introduction to and empirical application of autoencoder models in marketing
    Shankar, Venkatesh
    Parsana, Sohil
    JOURNAL OF THE ACADEMY OF MARKETING SCIENCE, 2022, 50 (06) : 1324 - 1350
  • [50] Customer satisfaction and natural language processing
    Piris, Yolande
    Gay, Anne-Cecile
    JOURNAL OF BUSINESS RESEARCH, 2021, 124 : 264 - 271