Improving Extraction of Chinese Open Relations Using Pre-trained Language Model and Knowledge Enhancement

被引:0
作者
Wen, Chaojie [1 ]
Jia, Xudong [1 ]
Chen, Tao [1 ]
机构
[1] WuYi Univ, Fac Intelligent Mfg, Jiangmen, Guangdong, Peoples R China
关键词
Chinese open relation extraction; Pre-trained language model; Knowledge enhancement; OPEN INFORMATION EXTRACTION;
D O I
10.1162/dint_a_00227
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Open Relation Extraction (ORE) is a task of extracting semantic relations from a text document. Current ORE systems have significantly improved their efficiency in obtaining Chinese relations, when compared with conventional systems which heavily depend on feature engineering or syntactic parsing. However, the ORE systems do not use robust neural networks such as pre-trained language models to take advantage of large-scale unstructured data effectively. In respons to this issue, a new system entitled Chinese Open Relation Extraction with Knowledge Enhancement (CORE-KE) is presented in this paper. The CORE-KE system employs a pre-trained language model (with the support of a Bidirectional Long Short-Term Memory (BiLSTM) layer and a Masked Conditional Random Field (Masked CRF) layer) on unstructured data in order to improve Chinese open relation extraction. Entity descriptions in Wikidata and additional knowledge (in terms of triple facts) extracted from Chinese ORE datasets are used to fine-tune the pre-trained language model. In addition, syntactic features are further adopted in the training stage of the CORE-KE system for knowledge enhancement. Experimental results of the CORE-KE system on two large-scale datasets of open Chinese entities and relations demonstrate that the CORE-KE system is superior to other ORE systems. The F1-scores of the CORE-KE system on the two datasets have given a relative improvement of 20.1% and 1.3%, when compared with benchmark ORE systems, respectively. The source code is available at https://github.com/cjwen15/CORE-KE.
引用
收藏
页码:962 / 989
页数:28
相关论文
共 51 条
  • [1] Angeli G, 2015, PROCEEDINGS OF THE 53RD ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 7TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 1, P344
  • [2] Cetto M., 2018, Graphene: Semantically-linked propositions in open information extraction, P2300
  • [3] Che WX, 2021, 2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021): PROCEEDINGS OF SYSTEM DEMONSTRATIONS, P42
  • [4] Chen ZY, 2020, IEEE INFOCOM SER, P307, DOI [10.1109/infocom41043.2020.9155531, 10.1109/INFOCOM41043.2020.9155531]
  • [5] Christensen J., 2011, An analysis of open information extraction based on semantic role labeling, P113
  • [6] ELECTRA: PRE-TRAINING TEXT ENCODERS AS DISCRIMINATORS RATHER THAN GENERATORS
    Clark, Kevin
    Luong, Minh-Thang
    Le, Quoc V.
    Manning, Christopher D.
    [J]. INFORMATION SYSTEMS RESEARCH, 2020,
  • [7] Cui L, 2018, PROCEEDINGS OF THE 56TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 2, P407
  • [8] Cui Y., 2020, ARXIV200413922, P657
  • [9] Pre-Training With Whole Word Masking for Chinese BERT
    Cui, Yiming
    Che, Wanxiang
    Liu, Ting
    Qin, Bing
    Yang, Ziqing
    [J]. IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2021, 29 : 3504 - 3514
  • [10] Del Corro Luciano, 2013, WWW