Hierarchical Question-Aware Context Learning with Augmented Data for Biomedical Question Answering

被引:0
作者
Du, Yongping [1 ]
Guo, Wenyang [1 ]
Zhao, Yiliang [1 ]
机构
[1] Beijing Univ Technol, Fac Informat Technol, Beijing, Peoples R China
来源
2019 IEEE INTERNATIONAL CONFERENCE ON BIOINFORMATICS AND BIOMEDICINE (BIBM) | 2019年
基金
国家重点研发计划;
关键词
biomedical question answering; domain adaptation; data augmentation; attention mechanism;
D O I
暂无
中图分类号
Q5 [生物化学];
学科分类号
071010 ; 081704 ;
摘要
This paper is concerned with the task of biomedical Question Answering (QA) which refers to extracting an answer to the given question from a biomedical context. Current works have made progress on this task, but they are still severely restricted by the insufficient training data due to the domain-specific nature, which motivates us to further explore a powerful way to solve this problem. We propose a Hierarchical Question-Aware Context Learning (HQACL) model for the biomedical QA task constituted by multi-level attention. The interaction between the question and the context can be captured layer by layer, with multi-grained embeddings to strengthen the ability of the language representation. A special training method called DA, including two parts namely domain adaptation and data augmentation, is also introduced to enhance the model performance. Domain adaptation can be defined as pre-training on a large-scale open-domain dataset and fine-tuning on the small training set of the target domain. As for the data augmentation, the Round-trip translation method is adopted to create new data with various expressions, which almost doubles the training set. The public biomedical dataset collected from PubMed provided by BioASQ is used to evaluate our model. The results show that our approach is superior to the best recent solution and achieves a new state of the art.
引用
收藏
页码:370 / 375
页数:6
相关论文
共 22 条
  • [1] Bahdanau Dzmitry, 2015, 3 INT C LEARN REPR I
  • [2] Devlin J., 2018, ARXIV
  • [3] Du YP, 2018, IEEE INT C BIOINFORM, P362, DOI 10.1109/BIBM.2018.8621197
  • [4] Fadaee M., 2018, ARXIV180809006
  • [5] Gehring J, 2017, PR MACH LEARN RES, V70
  • [6] A Pilot Study of Biomedical Text Comprehension using an Attention-Based Deep Neural Reader: Design and Experimental Analysis
    Kim, Seongsoon
    Park, Donghyeon
    Choi, Yonghwa
    Lee, Kyubum
    Kim, Byounggun
    Jeon, Minji
    Kim, Jihye
    Tan, Aik Choon
    Kang, Jaewoo
    [J]. JMIR MEDICAL INFORMATICS, 2018, 6 (01)
  • [7] Kingma DP, 2014, ARXIV
  • [8] Lee Jinhyuk, 2019, ABS190108746 CORR
  • [9] Lichtarge J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P3291
  • [10] Question Answering through Transfer Learning from Large Fine-grained Supervision Data
    Min, Sewon
    Seo, Minjoon
    Hajishirzi, Hannaneh
    [J]. PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 2, 2017, : 510 - 517