Generating Token-Level Explanations for Natural Language Inference

被引:0
|
作者
Thorne, James [1 ]
Vlachos, Andreas [1 ]
Christodoulopoulos, Christos [2 ]
Mittal, Arpit [2 ]
机构
[1] Univ Cambridge, Cambridge, England
[2] Amazon, Cambridge, England
来源
2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1 | 2019年
基金
欧盟地平线“2020”;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The task of Natural Language Inference (NLI) is widely modeled as supervised sentence pair classification. While there has been a lot of work recently on generating explanations of the predictions of classifiers on a single piece of text, there have been no attempts to generate explanations of classifiers operating on pairs of sentences. In this paper, we show that it is possible to generate token-level explanations for NLI without the need for training data explicitly annotated for this purpose. We use a simple LSTM architecture and evaluate both LIME and Anchor explanations for this task. We compare these to a Multiple Instance Learning (MIL) method that uses thresholded attention make token-level predictions. The approach we present in this paper is a novel extension of zero-shot single-sentence tagging to sentence pairs for NLI. We conduct our experiments on the well-studied SNLI dataset that was recently augmented with manually annotation of the tokens that explain the entailment relation. We find that our white-box MIL-based method, while orders of magnitude faster, does not reach the same accuracy as the black-box methods.
引用
收藏
页码:963 / 969
页数:7
相关论文
共 50 条
  • [41] cregit: Token-level blame information in git version control repositories
    Daniel M. German
    Bram Adams
    Kate Stewart
    Empirical Software Engineering, 2019, 24 : 2725 - 2763
  • [42] Toward Generating Natural-Language Explanations of Modal-Logic Proofs
    Giancola, Mike
    Bringsjord, Selmer
    Govindarajulu, Naveen Sundar
    ARTIFICIAL GENERAL INTELLIGENCE, AGI 2022, 2023, 13539 : 220 - 230
  • [43] Streaming Multi-Talker ASR with Token-Level Serialized Output Training
    Kanda, Naoyuki
    Wu, Jian
    Wu, Yu
    Xiao, Xiong
    Meng, Zhong
    Wang, Xiaofei
    Gaur, Yashesh
    Chen, Zhuo
    Li, Jinyu
    Yoshioka, Takuya
    INTERSPEECH 2022, 2022, : 3774 - 3778
  • [44] TokenMixup: Efficient Attention-guided Token-level Data Augmentation for Transformers
    Choi, Hyeong Kyu
    Choi, Joonmyung
    Kim, Hyunwoo J.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [45] Tuning Multi-mode Token-level Prompt Alignment across Modalities
    Wang, Dongsheng
    Li, Miaoge
    Liu, Xinyang
    Xu, MingSheng
    Chen, Bo
    Zhang, Hanwang
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [46] Injecting Wiktionary to improve token-level contextual representations using contrastive learning
    Mosolova, Anna
    Candito, Marie
    Ramisch, Carlos
    PROCEEDINGS OF THE 18TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 2: SHORT PAPERS, 2024, : 34 - 41
  • [47] Token-Level Self-Evolution Training for Sequence-to-Sequence Learning
    Peng, Keqin
    Ding, Liang
    Zhong, Qihuang
    Ouyang, Yuanxin
    Rong, Wenge
    Xiong, Zhang
    Tao, Dacheng
    61ST CONFERENCE OF THE THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 2, 2023, : 841 - 850
  • [48] Incorporating token-level dictionary feature into neural model for named entity recognition
    Mu Xiaofeng
    Wang Wei
    Xu Aiping
    NEUROCOMPUTING, 2020, 375 : 43 - 50
  • [49] Generating Data to Mitigate Spurious Correlations in Natural Language Inference Datasets
    Wu, Yuxiang
    Gardner, Matt
    Stenetorp, Pontus
    Dasigi, Pradeep
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 2660 - 2676
  • [50] Generating Natural Counterfactual Visual Explanations
    Zhao, Wenqi
    Oyama, Satoshi
    Kurihara, Masahito
    PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 5204 - 5205