An Empirical Study of Multi-Task Learning on BERT for Biomedical Text Mining

被引:0
|
作者
Peng, Yifan [1 ]
Chen, Qingyu [1 ]
Lu, Zhiyong [1 ]
机构
[1] NIH, Natl Ctr Biotechnol Informat, Natl Lib Med, Bldg 10, Bethesda, MD 20892 USA
来源
19TH SIGBIOMED WORKSHOP ON BIOMEDICAL LANGUAGE PROCESSING (BIONLP 2020) | 2020年
基金
美国国家卫生研究院;
关键词
CORPUS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-task learning (MTh) has achieved remarkable success in natural language processing applications. In this work, we study a multi-task learning model with multiple decoders on varieties of biomedical and clinical natural language processing tasks such as text similarity, relation extraction, named entity recognition, and text inference. Our empirical results demonstrate that the MTL fine-tuned models outperform state-of-the-art transformer models (e.g., BERT and its variants) by 2.0% and 1.3% in biomedical and clinical domains, respectively. Pairwise MTL further demonstrates more details about which tasks can improve or decrease others. This is particularly helpful in the context that researchers are in the hassle of choosing a suitable model for new problems. The code and models are publicly available at https://github.com/ncbi-nlp/bluebert.
引用
收藏
页码:205 / 214
页数:10
相关论文
共 50 条
  • [21] Leveraging Multi-task Learning for Biomedical Named Entity Recognition
    Mehmood, Tahir
    Gerevini, Alfonso
    Lavelli, Alberto
    Serina, Ivan
    ADVANCES IN ARTIFICIAL INTELLIGENCE, AI*IA 2019, 2019, 11946 : 431 - 444
  • [22] Biomedical Named Entity Recognition Based on Multi-task Learning
    Zhao, Hui
    Zhao, Di
    Meng, Jiana
    Su, Wen
    Mu, Wenxuan
    HEALTH INFORMATION PROCESSING, CHIP 2023, 2023, 1993 : 51 - 65
  • [23] Multi-task transfer learning for biomedical machine reading comprehension
    Guo, Wenyang
    Du, Yongping
    Zhao, Yiliang
    Ren, Keyan
    INTERNATIONAL JOURNAL OF DATA MINING AND BIOINFORMATICS, 2020, 23 (03) : 234 - 250
  • [24] Multi-task gradient descent for multi-task learning
    Lu Bai
    Yew-Soon Ong
    Tiantian He
    Abhishek Gupta
    Memetic Computing, 2020, 12 : 355 - 369
  • [25] Multi-task gradient descent for multi-task learning
    Bai, Lu
    Ong, Yew-Soon
    He, Tiantian
    Gupta, Abhishek
    MEMETIC COMPUTING, 2020, 12 (04) : 355 - 369
  • [26] BERT and PALs: Projected Attention Layers for Efficient Adaptation in Multi-Task Learning
    Stickland, Asa Cooper
    Murray, Iain
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [27] Multi-task prediction method of business process based on BERT and Transfer Learning
    Chen, Hang
    Fang, Xianwen
    Fang, Huan
    KNOWLEDGE-BASED SYSTEMS, 2022, 254
  • [28] Multi-Task Learning Using BERT With Soft Parameter Sharing Between Layers
    Pahari, Niraj
    Shimada, Kazutaka
    2022 JOINT 12TH INTERNATIONAL CONFERENCE ON SOFT COMPUTING AND INTELLIGENT SYSTEMS AND 23RD INTERNATIONAL SYMPOSIUM ON ADVANCED INTELLIGENT SYSTEMS (SCIS&ISIS), 2022,
  • [29] Ask the GRU: Multi-task Learning for Deep Text Recommendations
    Bansal, Trapit
    Belanger, David
    McCallum, Andrew
    PROCEEDINGS OF THE 10TH ACM CONFERENCE ON RECOMMENDER SYSTEMS (RECSYS'16), 2016, : 107 - 114
  • [30] Multi-task Learning with Bidirectional Language Models for Text Classification
    Yang, Qi
    Shang, Lin
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,