Deep Entity Linking via Eliminating Semantic Ambiguity With BERT

被引:18
作者
Yin, Xiaoyao [1 ]
Huang, Yangchen [1 ]
Zhou, Bin [1 ]
Li, Aiping [1 ]
Lan, Long [1 ,2 ]
Jia, Yan [1 ]
机构
[1] Natl Univ Def Technol, Coll Comp, Changsha 410073, Peoples R China
[2] Natl Univ Def Technol, State Key Lab High Performance Comp, Changsha 410073, Peoples R China
来源
IEEE ACCESS | 2019年 / 7卷
基金
中国国家自然科学基金;
关键词
Task analysis; Knowledge based systems; Semantics; Joining processes; Bit error rate; Natural languages; Computational modeling; Entity linking; natural language processing (NLP); bidirectional encoder representations from transformers (BERT); deep neural network (DNN);
D O I
10.1109/ACCESS.2019.2955498
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Entity linking refers to the task of aligning mentions of entities in the text to their corresponding entries in a specific knowledge base, which is of great significance for many natural language process applications such as semantic text understanding and knowledge fusion. The pivotal of this problem is how to make effective use of contextual information to disambiguate mentions. Moreover, it has been observed that, in most cases, mention has similar or even identical strings to the entity it refers to. To prevent the model from linking mentions to entities with similar strings rather than the semantically similar ones, in this paper, we introduce the advanced language representation model called BERT (Bidirectional Encoder Representations from Transformers) and design a hard negative samples mining strategy to fine-tune it accordingly. Based on the learned features, we obtain the valid entity through computing the similarity between the textual clues of mentions and the entity candidates in the knowledge base. The proposed hard negative samples mining strategy benefits entity linking from the larger, more expressive pre-trained representations of BERT with limited training time and computing sources. To the best of our knowledge, we are the first to equip entity linking task with the powerful pre-trained general language model by deliberately tackling its potential shortcoming of learning literally, and the experiments on the standard benchmark datasets show that the proposed model yields state-of-the-art results.
引用
收藏
页码:169434 / 169445
页数:12
相关论文
共 44 条
  • [41] Extracting Higher Order Topological Semantic via Motif-Based Deep Graph Neural Networks
    Zhang, Ke-Jia
    Ding, Xiao
    Xiang, Bing-Bing
    Zhang, Hai-Feng
    Bao, Zhong-Kui
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2024, 11 (04) : 5444 - 5453
  • [42] Semantic analysis via application of deep learning using Naver movie review data
    Kim, Sojin
    Song, Jongwoo
    KOREAN JOURNAL OF APPLIED STATISTICS, 2022, 35 (01) : 19 - 33
  • [43] A Novel Approach To Populate Multimedia Knowledge Graph Via Deep Learning And Semantic Analysis
    Rinaldi, Antonio M.
    Russo, Cristiano
    Tommasino, Cristian
    PROCEEDINGS OF 2022 14TH INTERNATIONAL CONFERENCE ON MANAGEMENT OF DIGITAL ECOSYSTEMS, MEDES 2022, 2022, : 40 - 47
  • [44] A Semantic-Embedding Model-Driven Seq2Seq Method for Domain-Oriented Entity Linking on Resource-Restricted Devices
    Inan, Emrah
    Dikenelli, Oguz
    INTERNATIONAL JOURNAL ON SEMANTIC WEB AND INFORMATION SYSTEMS, 2021, 17 (03) : 73 - 87