Improving biomedical Named Entity Recognition with additional external contexts

被引:0
作者
Tho, Bui Duc [1 ,3 ]
Nguyen, Minh -Tien [1 ]
Le, Dung Tien [2 ]
Ying, Lin -Lung [2 ]
Inoue, Shumpei [2 ]
Nguyen, Tri-Thanh [3 ]
机构
[1] Hung Yen Univ Technol & Educ, Hung Yen, Vietnam
[2] CINNAMON LAB, 10th Floor,Geleximco Bldg,36 Hoang Cau, Hanoi, Vietnam
[3] Vietnam Natl Univ, Univ Engn & Technol, Hanoi, Vietnam
关键词
Biomedical Named Entity Recognition; Information extraction; Transformers; External contexts;
D O I
10.1016/j.jbi.2024.104674
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Objective: Biomedical Named Entity Recognition (bio NER) is the task of recognizing named entities in biomedical texts. This paper introduces a new model that addresses bio NER by considering additional external contexts. Different from prior methods that mainly use original input sequences for sequence labeling, the model takes into account additional contexts to enhance the representation of entities in the original sequences, since additional contexts can provide enhanced information for the concept explanation of biomedical entities. Methods: To exploit an additional context, given an original input sequence, the model first retrieves the relevant sentences from PubMed and then ranks the retrieved sentences to form the contexts. It next combines the context with the original input sequence to form a new enhanced sequence. The original and new enhanced sequences are fed into PubMedBERT for learning feature representation. To obtain more fine-grained features, the model stacks a BiLSTM layer on top of PubMedBERT. The final named entity label prediction is done by using a CRF layer. The model is jointly trained in an end -to -end manner to take advantage of the additional context for NER of the original sequence. Results: Experimental results on six biomedical datasets show that the proposed model achieves promising performance compared to strong baselines and confirms the contribution of additional contexts for bio NER. Conclusion: The promising results confirm three important points. First, the additional context from PubMed helps to improve the quality of the recognition of biomedical entities. Second, PubMed is more appropriate than the Google search engine for providing relevant information of bio NER. Finally, more relevant sentences from the context are more beneficial than irrelevant ones to provide enhanced information for the original input sequences. The model is flexible to integrate any additional context types for the NER task.
引用
收藏
页数:11
相关论文
共 65 条
[1]  
Agarwal O, 2021, FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, P3990
[2]  
Alsentzer Emily, 2019, P 2 CLIN NAT LANG PR, P72, DOI [DOI 10.18653/V1/W19-1909, 10.18653/v1/W19-1909]
[3]  
Amatriain X, 2024, Arxiv, DOI [arXiv:2302.07730, DOI 10.48550/ARXIV.2302.07730, 10.48550/arXiv.2302.07730]
[4]  
Angeli G, 2015, PROCEEDINGS OF THE 53RD ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 7TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 1, P344
[5]  
[Anonymous], 2007, P 2007 JOINT C EMP M
[6]  
[Anonymous], 2016, NAACL HLT 2016 2016, DOI [DOI 10.18653/V1/N16-1030, 10 . 18653 / v1 / N16-1030]
[7]  
[Anonymous], 2015, EMNLP
[8]  
Banerjee P, 2020, Arxiv, DOI arXiv:1911.03869
[9]  
Beltagy I, 2019, 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019), P3615
[10]  
Brown TB, 2020, ADV NEUR IN, V33