Continuous Prompt Enhanced Biomedical Entity Normalization

被引:2
作者
Lai, Zhaohong [1 ,2 ]
Fu, Biao [1 ,2 ]
Wei, Shangfei [1 ,2 ]
Shi, Xiaodong [1 ,2 ]
机构
[1] Xiamen Univ, Sch Informat, Dept Artificial Intelligence, Xiamen, Peoples R China
[2] Minist Culture & Tourism, Key Lab Digital Protect & Intelligent Proc Intang, Xiamen, Peoples R China
来源
NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, NLPCC 2022, PT II | 2022年 / 13552卷
关键词
Prompt-BEN; Prompt learning; Contrastive loss; RECOGNITION;
D O I
10.1007/978-3-031-17189-5_5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Biomedical entity normalization (BEN) aims to link the entity mentions in a biomedical text to referent entities in a knowledge base. Recently, the paradigm of large-scale language model pre-training and fine-tuning have achieved superior performance in BEN task. However, pre-trained language models like SAPBERT [21] typically contain hundreds of millions of parameters, and fine-tuning all parameters is computationally expensive. The latest research such as prompt technology is proposed to reduce the amount of parameters during the model training. Therefore, we propose a framework Prompt-BEN using continuous Prompt to enhance BEN, which just needs to fine-tune few parameters of prompt. Our method employs embeddings with the continuous prefix prompt to capture the semantic similarity between mention and terms. We also design a contrastive loss with synonym marginalized strategy for the BEN task. Finally, experimental results on three benchmark datasets demonstrated that our method achieves competitive or even greater linking accuracy than the state-of-the-art fine-tuning-based models while having about 600 times fewer tuned parameters.
引用
收藏
页码:61 / 72
页数:12
相关论文
共 39 条
  • [1] Alsentzer E., 2019, 2 CLIN NATURAL LANG, DOI [DOI 10.18653/V1/W19-1909, 10.18653/v1/W19-1909]
  • [2] The Unified Medical Language System (UMLS): integrating biomedical terminology
    Bodenreider, O
    [J]. NUCLEIC ACIDS RESEARCH, 2004, 32 : D267 - D270
  • [3] Chen L., 2020, ARXIV
  • [4] D'Souza J, 2015, PROCEEDINGS OF THE 53RD ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL) AND THE 7TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (IJCNLP), VOL 2, P297
  • [5] Dogan R., 2012, AAAI FALL S
  • [6] NCBI disease corpus: A resource for disease name recognition and concept normalization
    Dogan, Rezarta Islamaj
    Leaman, Robert
    Lu, Zhiyong
    [J]. JOURNAL OF BIOMEDICAL INFORMATICS, 2014, 47 : 1 - 10
  • [7] Bringing Transparency Design into Practice
    Eiband, Malin
    Schneider, Hanna
    Bilandzic, Mark
    Fazekas-Con, Julian
    Haug, Mareike
    Hussmann, Heinrich
    [J]. IUI 2018: PROCEEDINGS OF THE 23RD INTERNATIONAL CONFERENCE ON INTELLIGENT USER INTERFACES, 2018, : 211 - 223
  • [8] NSEEN: Neural Semantic Embedding for Entity Normalization
    Fakhraei, Shobeir
    Mathew, Joel
    Ambite, Jose Luis
    [J]. MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2019, PT II, 2020, 11907 : 665 - 680
  • [9] Hadsell R., 2006, CVPR, P1735
  • [10] Ji Zongcheng, 2020, AMIA Jt Summits Transl Sci Proc, V2020, P269