Multi-task transfer learning for biomedical machine reading comprehension

被引:0
作者
Guo, Wenyang [1 ]
Du, Yongping [1 ]
Zhao, Yiliang [1 ]
Ren, Keyan [1 ]
机构
[1] Beijing Univ Technol, Fac Informat Technol, Beijing 100124, Peoples R China
基金
国家重点研发计划;
关键词
biomedical machine reading comprehension; multi-task learning; transfer learning; attention; data augmentation;
D O I
10.1504/IJDMB.2020.107878
中图分类号
Q [生物科学];
学科分类号
07 ; 0710 ; 09 ;
摘要
Biomedical machine reading comprehension aims to extract the answer to the given question from complex biomedical passages, which requires the machine to have the ability to process strong comprehension on natural language. Recent progress has made on this task, but still severely restricted by the insufficient training data due to the domain-specific nature. To solve this problem, we propose a hierarchical question-aware context learning model trained by the multi-task transfer learning algorithm, which can capture the interaction between the question and the passage layer by layer, with multi-level embeddings to strengthen the ability of the language representation. The multi-task transfer learning algorithm leverages the advantages of different machine reading comprehension tasks to improve model generalisation and robustness, pre-training on multiple large-scale open-domain data sets and fine-tuning on the target-domain training set. Moreover, data augmentation is also adopted to create new training samples with various expressions. The public biomedical data set collected from PubMed provided by BioASQ is used to evaluate the model performance. The results show that our method is superior to the best recent solution and achieves a new state of the art.
引用
收藏
页码:234 / 250
页数:17
相关论文
共 29 条
[1]  
Bahdanau D, 2016, Arxiv, DOI [arXiv:1409.0473, DOI 10.48550/ARXIV.1409.0473]
[2]  
Cheng Y., 2015, P 25 INT JOINT C ART, P2761
[3]  
Chung Y.-A., 2018, P 2018 C N AM CHAPTE, V1, P1585, DOI [10.18653/v1/N18-1143, DOI 10.18653/V1, DOI 10.18653/V1/N18-1143]
[4]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[5]  
Hu MH, 2018, 2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), P2077
[6]  
Jia R., 2017, P 2017 C EMP METH NA, P2011
[7]   TriviaQA: A Large Scale Distantly Supervised Challenge Dataset for Reading Comprehension [J].
Joshi, Mandar ;
Choi, Eunsol ;
Weld, Daniel S. ;
Zettlemoyer, Luke .
PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 1, 2017, :1601-1611
[8]  
Kim SY, 2014, J ADV PROSTHODONT, V6, P1
[9]   BioBERT: a pre-trained biomedical language representation model for biomedical text mining [J].
Lee, Jinhyuk ;
Yoon, Wonjin ;
Kim, Sungdong ;
Kim, Donghyeon ;
Kim, Sunkyu ;
So, Chan Ho ;
Kang, Jaewoo .
BIOINFORMATICS, 2020, 36 (04) :1234-1240
[10]  
Lichtarge J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P3291