Chinese Triple Extraction Based on BERT Model

被引:2
|
作者
Deng, Weidong [1 ,2 ]
Liu, Yun [1 ,2 ]
机构
[1] Beijing Jiao Tong Univ, Sch Elect & Informat Engn, Beijing, Peoples R China
[2] Beijing Municipal Commiss Educ, Key Lab Commun & Informat Syst, Beijing, Peoples R China
来源
PROCEEDINGS OF THE 2021 15TH INTERNATIONAL CONFERENCE ON UBIQUITOUS INFORMATION MANAGEMENT AND COMMUNICATION (IMCOM 2021) | 2021年
基金
中国国家自然科学基金;
关键词
triple; information extraction; relation classification; entity tagging; BERT;
D O I
10.1109/IMCOM51814.2021.9377404
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Information extraction (IE) plays a crucial role in natural language processing, which extracts structured facts like entities, attributes, relations and events from unstructured text. The results of information extraction can be applied in many fields including information retrieval, intelligent QA system, to name a few. We define a pair of entities and their relation from a sentence as a triple. Different from most relation extraction tasks, which only extract one relation from a sentence of known entities, we achieved that extracting both relation and entities(a triple, as defined above), from a plain sentence. Until now, there are so many methods proposed to solve information extraction problem and deep learning has made great progress last several years. Among the field of deep learning, the pretrained model BERT has achieved greatly successful results in a lot of NLP tasks. So we divide our triple extraction task into two sub-tasks, relation classification and entity tagging, and design two models based on BERT for these two sub-tasks, including a CNN-BERT and a Simple BERT. We experimented our models on DuIE Chinese dataset and achieved excellent results.
引用
收藏
页数:5
相关论文
共 50 条
  • [31] AMP-BERT: Prediction of antimicrobial peptide function based on a BERT model
    Lee, Hansol
    Lee, Songyeon
    Lee, Ingoo
    Nam, Hojung
    PROTEIN SCIENCE, 2023, 32 (01)
  • [32] Chinese Text Classification Method Based on BERT Word Embedding
    Wang, Ziniu
    Huang, Zhilin
    Gao, Jianling
    2020 5TH INTERNATIONAL CONFERENCE ON MATHEMATICS AND ARTIFICIAL INTELLIGENCE (ICMAI 2020), 2020, : 66 - 71
  • [33] Enhancing Sentiment Analysis for Chinese Texts Using a BERT-Based Model with a Custom Attention Mechanism
    Ding, Linlin
    Han, Yiming
    Li, Mo
    Li, Dong
    WEB INFORMATION SYSTEMS AND APPLICATIONS, WISA 2024, 2024, 14883 : 172 - 179
  • [34] Constructing Chinese Historical Literature Knowledge Graph Based on BERT
    Guo, Qingyan
    Sun, Yang
    Liu, Guanzhong
    Wang, Zijun
    Ji, Zijing
    Shen, Yuxin
    Wang, Xin
    WEB INFORMATION SYSTEMS AND APPLICATIONS (WISA 2021), 2021, 12999 : 323 - 334
  • [35] Stacking-BERT model for Chinese medical procedure entity normalization
    Li, Luqi
    Zhai, Yunkai
    Gao, Jinghong
    Wang, Linlin
    Hou, Li
    Zhao, Jie
    MATHEMATICAL BIOSCIENCES AND ENGINEERING, 2023, 20 (01) : 1018 - 1036
  • [36] A BERT-based Idiom Detection Model
    Gamage, Gihan
    De Silva, Daswin
    Adikari, Achini
    Alahakoon, Damminda
    2022 15TH INTERNATIONAL CONFERENCE ON HUMAN SYSTEM INTERACTION (HSI), 2022,
  • [37] Emerging industry classification based on BERT model
    Yang, Baocheng
    Zhang, Bing
    Cutsforth, Kevin
    Yu, Shanfu
    Yu, Xiaowen
    INFORMATION SYSTEMS, 2025, 128
  • [38] Information Extraction from Chinese Papers Based on Hidden Markov Model
    Chi Chengying
    Zhang Yan
    ADVANCES IN MECHATRONICS, AUTOMATION AND APPLIED INFORMATION TECHNOLOGIES, PTS 1 AND 2, 2014, 846-847 : 1291 - 1294
  • [39] RESEARCH ON EVENT EXTRACTION MODEL BASED ON SEMANTIC FEATURES OF CHINESE WORDS
    Zhu, Shaowu
    Sun, Haichun
    Jian, Hanying
    COMPUTING AND INFORMATICS, 2022, 41 (06) : 1625 - 1647
  • [40] An enhanced guided LDA model augmented with BERT based semantic strength for aspect term extraction in sentiment analysis
    Venugopalan, Manju
    Gupta, Deepa
    KNOWLEDGE-BASED SYSTEMS, 2022, 246