KEPT: Knowledge Enhanced Prompt Tuning for event causality identification

被引:28
作者
Liu, Jintao [1 ,2 ,3 ,4 ]
Zhang, Zequn [1 ,2 ]
Guo, Zhi [1 ,2 ]
Jin, Li [1 ,2 ]
Li, Xiaoyu [1 ,2 ]
Wei, Kaiwen [1 ,2 ,3 ,4 ]
Sun, Xian [1 ,2 ,3 ]
机构
[1] Chinese Acad Sci, Aerosp Informat Res Inst, Beijing 100190, Peoples R China
[2] Chinese Acad Sci, Aerosp Informat Res Inst, Key Lab Network Informat Syst Technol NIST, Beijing 100190, Peoples R China
[3] Univ Chinese Acad Sci, Beijing 100190, Peoples R China
[4] Univ Chinese Acad Sci, Sch Elect Elect & Commun Engn, Beijing 100190, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Event causality identification; Prompt tuning; Attention mechanism; Knowledge representation learning; MODEL;
D O I
10.1016/j.knosys.2022.110064
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Event causality identification (ECI) aims to identify causal relations of event mention pairs in text. Despite achieving certain accomplishments, existing methods are still not effective due to the following two issues: (1) the lack of causal reasoning ability, imposing restrictions on recognizing implicit causal relations; (2) the significant gap between fine-tuning and pre-training, which hinders the utilization of pre-trained language models (PLMs). In this paper, we propose a novel Knowledge Enhanced Prompt Tuning (KEPT) framework for ECI to address the issues mentioned above. Specifically, this method leverages prompt tuning to incorporate two kinds of knowledge obtained from external knowledge bases (KBs), including background information and relational information, for causal reasoning. To introduce external knowledge into our model, we first convert it to textual descriptions, then design an interactive attention mechanism and a selective attention mechanism to fuse background information and relational information, respectively. In addition, to further capture implicit relations between events, we adopt the objective from knowledge representation learning to jointly optimize the representations of causal relations and events. Experiment results on two widely-used benchmarks demonstrate that the proposed method outperforms the state-of-the-art models.(c) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页数:12
相关论文
共 74 条
[1]  
Beamer B, 2009, LECT NOTES COMPUT SC, V5449, P430, DOI 10.1007/978-3-642-00382-0_35
[2]  
Berant J., 2014, P 2014 C EMP METH NA, P1499, DOI [10.3115/v1/D14-1159, DOI 10.3115/V1/D14-1159]
[3]  
Bordes Antoine, 2013, PROCADV NEURAL INF P, V26
[4]  
Brown TB, 2020, ADV NEUR IN, V33
[5]  
Cao PF, 2021, 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, P4862
[6]  
Caselli T., 2017, P EV STOR NEWS WORKS, P77, DOI 10.18653/v1/W17-2711
[7]  
Chen X, 2021, Arxiv, DOI arXiv:2109.00720
[8]  
Chen X, 2023, Arxiv, DOI arXiv:2104.07650
[9]   KnowPrompt: Knowledge-aware Prompt-tuning with Synergistic Optimization for Relation Extraction [J].
Chen, Xiang ;
Zhang, Ningyu ;
Xie, Xin ;
Deng, Shumin ;
Yao, Yunzhi ;
Tan, Chuanqi ;
Huang, Fei ;
Si, Luo ;
Chen, Huajun .
PROCEEDINGS OF THE ACM WEB CONFERENCE 2022 (WWW'22), 2022, :2778-2788
[10]   Classifying Temporal Relations by Bidirectional LSTM over Dependency Paths [J].
Cheng, Fei ;
Miyao, Yusuke .
PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 2, 2017, :1-6