StaResGRU-CNN with CMedLMs: A stacked residual GRU-CNN with pre-trained biomedical language models for predictive intelligence

被引:17
作者
Ni, Pin [1 ]
Li, Gangmin [2 ]
Hung, Patrick C. K. [3 ]
Chang, Victor [4 ]
机构
[1] UCL, Sch Engn, London, England
[2] Univ Bedfordshire, Sch Comp Sci & Technol, Luton, Beds, England
[3] Ontario Tech Univ, Fac Business & Informat Technol, Oshawa, ON, Canada
[4] Teesside Univ, Sch Comp Engn & Digital Technol, Artificial Intelligence & Informat Syst Res Grp, Middlesbrough, Cleveland, England
关键词
Natural language processing; Predictive intelligence; Biomedical text mining; Named Entity Recognition; Text classification; Transfer learning; Pre-trained language model; SENTIMENT ANALYSIS; NEURAL-NETWORK; TEXT;
D O I
10.1016/j.asoc.2021.107975
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
As a task requiring strong professional experience as supports, predictive biomedical intelligence cannot be separated from the support of a large amount of external domain knowledge. By using transfer learning to obtain sufficient prior experience from massive biomedical text data, it is essential to promote the performance of specific downstream predictive and decision-making task models. This is an efficient and convenient method, but it has not been fully developed for Chinese Natural Language Processing (NLP) in the biomedical field. This study proposes a Stacked Residual Gated Recurrent Unit-Convolutional Neural Networks (StaResGRU-CNN) combined with the pre-trained language models (PLMs) for biomedical text-based predictive tasks. Exploring related paradigms in biomedical NLP based on transfer learning of external expert knowledge and comparing some Chinese and English language models. We have identified some key issues that have not been developed or those present difficulties of application in the field of Chinese biomedicine. Therefore, we also propose a series of Chinese bioMedical Language Models (CMedLMs) with detailed evaluations of downstream tasks. By using transfer learning, language models are introduced with prior knowledge to improve the performance of downstream tasks and solve specific predictive NLP tasks related to the Chinese biomedical field to serve the predictive medical system better. Additionally, a free-form text Electronic Medical Record (EMR)-based Disease Diagnosis Prediction task is proposed, which is used in the evaluation of the analyzed language models together with Clinical Named Entity Recognition, Biomedical Text Classification tasks. Our experiments prove that the introduction of biomedical knowledge in the analyzed models significantly improves their performance in the predictive biomedical NLP tasks with different granularity. And our proposed model also achieved competitive performance in these predictive intelligence tasks. (C) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页数:14
相关论文
共 69 条
  • [1] Alsentzer E., 2019, P 2 CLIN NAT LANG PR, P72, DOI [10.18653/v1/W19-1909, DOI 10.18653/V1/W19-1909]
  • [2] Predictive intelligence to the edge through approximate collaborative context reasoning
    Anagnostopoulos, Christos
    Kolomvatsos, Kostas
    [J]. APPLIED INTELLIGENCE, 2018, 48 (04) : 966 - 991
  • [3] Enhancing deep learning sentiment analysis with ensemble techniques in social applications
    Araque, Oscar
    Corcuera-Platas, Ignacio
    Sanchez-Rada, J. Fernando
    Iglesias, Carlos A.
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2017, 77 : 236 - 246
  • [4] Bharadwaj A., 2016, P 2016 C EMPIRICAL M, V2016, P1462, DOI DOI 10.18653/V1/D16-1153
  • [5] Random forests
    Breiman, L
    [J]. MACHINE LEARNING, 2001, 45 (01) : 5 - 32
  • [6] Chalkidis I, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), P6314
  • [7] From Code to Bedside: Introducing Predictive Intelligence to Interventional Oncology
    Chapiro, Julius
    Duncan, James S.
    [J]. RADIOLOGY-ARTIFICIAL INTELLIGENCE, 2019, 1 (05)
  • [8] Chatterjee Rajen, 2017, WMT 2017, V1, P157
  • [9] Chen Q, 2018, PROCEEDINGS OF THE 56TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL), VOL 1, P2406
  • [10] Chen Qian, 2017, ARXIV171104289