Self-supervised commonsense knowledge learning for document-level relation extraction

被引:0
作者
Li, Rongzhen [1 ,2 ]
Zhong, Jiang [1 ,2 ]
Xue, Zhongxuan [1 ]
Dai, Qizhu [1 ]
Li, Xue [3 ]
机构
[1] Chongqing Univ, Coll Comp Sci, Chongqing 400044, Peoples R China
[2] Chongqing Univ, Key Lab Dependable Serv Comp Cyber Phys Soc, Minist Educ, Chongqing 400044, Peoples R China
[3] Univ Queensland, Sch Informat Technol & Elect Engn, Brisbane, Australia
基金
中国国家自然科学基金;
关键词
Document-level relation extraction; Self-supervised learning; Commonsense knowledge;
D O I
10.1016/j.eswa.2024.123921
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Compared to sentence -level relation extraction, practical document -level relation extraction (DocRE) is a more challenging task for which multi -entity problems need to be resolved. It aims at extracting relationships between two entities over multiple sentences at once while taking into account significant cross -sentence features. Learning long-distance semantic relation representation across sentences in a document, however, is a widespread and difficult task. To address the above issues, we propose a novel S elf -supervised C ommonsense - enhanced D oc RE approach, named as SCDRE , bypassing the need for external knowledge. The methodology begins by harnessing self -supervised learning to capture the commonsense knowledge pertaining to each entity within an entity pair, drawing insights from the commonsense entailed text. This acquired knowledge subsequently serves as the foundation for transforming cross -sentence entity pairs into alias counterparts achieved by the coreference commonsense replacement. The focus then shifts to semantic relation representation learning, applied to these alias entity pairs. Through a process of entity pair rich attention fusion, these alias pairs are seamlessly and automatically translated back into the target entity pairs. This innovation harnesses self -supervised learning and contextual commonsense to distinguish SCDRE as a unique and selfcontained approach, promising an enhanced ability to extract relationships from documents. We examine our model on three publicly accessible datasets, DocRED, DialogRE and MPDD, and the results show that it performs significantly better than strong baselines by 2.03% F1, and commonsense knowledge has an important contribution to the DocRE by the ablation experimental analysis.
引用
收藏
页数:9
相关论文
共 54 条
[51]   Attention-Based Bidirectional Long Short-Term Memory Networks for Relation Classification [J].
Zhou, Peng ;
Shi, Wei ;
Tian, Jun ;
Qi, Zhenyu ;
Li, Bingchen ;
Hao, Hongwei ;
Xu, Bo .
PROCEEDINGS OF THE 54TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2016), VOL 2, 2016, :207-212
[52]  
Zhou W., 2020, INT C LEARN REPR
[53]  
Zhou WX, 2021, AAAI CONF ARTIF INTE, V35, P14612
[54]  
Zhu H, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), P1331