TERL: Transformer Enhanced Reinforcement Learning for Relation Extraction

被引:0
作者
Wang, Yashen [1 ,2 ]
Shi, Tuo [3 ]
Ouyang, Xiaoye [1 ]
Guo, Dayu [4 ]
机构
[1] China Acad Elect & Informat Technol, Natl Engn Lab Risk Percept & Prevent RPP, Beijing 100041, Peoples R China
[2] CETC, Artificial Intelligence Inst, Key Lab Cognit & Intelligence Technol CIT, Beijing 100144, Peoples R China
[3] Beijing Police Coll, Beijing 102202, Peoples R China
[4] CETC Acad Elect & Informat Technol Grp Co Ltd, Beijing 100041, Peoples R China
来源
CHINESE COMPUTATIONAL LINGUISTICS, CCL 2023 | 2023年 / 14232卷
基金
中国国家自然科学基金;
关键词
Relation Extraction; Reinforcement Learning; Transformer;
D O I
10.1007/978-981-99-6207-5_12
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Relation Extraction (RE) task aims to discover the semantic relation that holds between two entities and contributes to many applications such as knowledge graph construction and completion. Reinforcement Learning (RL) has been widely used for RE task and achieved SOTA results, which are mainly designed with rewards to choose the optimal actions during the training procedure, to improve RE's performance, especially for low-resource conditions. Recent work has shown that offline or online RL can be flexibly formulated as a sequence understanding problem and solved via approaches similar to large-scale pre-training language modeling. To strengthen the ability for understanding the semantic signals interactions among the given text sequence, this paper leverages Transformer architecture for RL-based RE methods, and proposes a generic framework called Transformer Enhanced RL (TERL) towards RE task. Unlike prior RL-based RE approaches that usually fit value functions or compute policy gradients, TERL only outputs the best actions by utilizing a masked Transformer. Experimental results show that the proposed TERL framework can improve many state-of-the-art RL-based RE methods.
引用
收藏
页码:192 / 206
页数:15
相关论文
共 46 条
[1]  
Chen Lili, 2021, NEURIPS
[2]  
Chen Yi, 2021, EMNLP
[3]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[4]   Creating Training Corpora for NLG Micro-Planning [J].
Gardent, Claire ;
Shimorina, Anastasia ;
Narayan, Shashi ;
Perez-Beltrachini, Laura .
PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 1, 2017, :179-188
[5]  
Gormley M. R., 2015, P C EMP METH NAT LAN, P1774
[6]   A Deep Relevance Matching Model for Ad-hoc Retrieval [J].
Guo, Jiafeng ;
Fan, Yixing ;
Ai, Qingyao ;
Croft, W. Bruce .
CIKM'16: PROCEEDINGS OF THE 2016 ACM CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, 2016, :55-64
[7]  
Guo ZJ, 2020, PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P3651
[8]  
Hendrickx Iris, 2010, P 5 INT WORKSH SEM E, P33, DOI DOI 10.48550/ARXIV.1911.10422
[9]  
Hochreiter S, 1997, NEURAL COMPUT, V9, P1735, DOI [10.1162/neco.1997.9.1.1, 10.1007/978-3-642-24797-2]
[10]  
Hoffmann R., 2011, ANN M ASS COMP LING, P541