Design and Implementation of Information Extraction System for Scientific Literature Using Fine-tuned Deep Learning Models

被引:0
作者
Won, Kwanghee [1 ]
Jang, Youngsun [1 ]
Choi, Hyung-do [2 ]
Shin, Sung [1 ]
机构
[1] South Dakota State Univ, Comp Sci, Brookings, SD 57007 USA
[2] Elect & Telecom Res Inst, Daejeon, South Korea
来源
APPLIED COMPUTING REVIEW | 2022年 / 22卷 / 01期
关键词
Deep Learning; Question Answering; Semantic Classification; Bidirectional Encoder Representations of Transformers (BERT);
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper presents an overview of a quality scoring system that utilizes pre-trained deep neural network models. Two types of DL models, a classification and extractive question answering (EQA) models are used to implement components of the system. The abstracts of the scientific literature are classified into two groups, in-vivo and in-vitro, and a question and answering model architecture is constructed for extracting the following types of information (animal type, the number of animals, exposure dose, and signal frequency). The Bidirectional Encoder Representations of Transformers (BERT) model pre-trained with a large text corpus is used as our baseline model for classification and EQA tasks. The models are fine-tuned with 455 EMF-related research papers. In our experiments, the fine-tuned model showed improved performance on EQA tasks for the four-categories of questions compared to the baseline, and it also showed improvements on similar questions that were not used in training. This suggests the importance of retraining of deep learning model specifically in some areas requiring domain expertise such as scientific papers. However, additional research is needed on some implementation issues, in such cases where there are still multiple answers, or where there is no answer given in a context.
引用
收藏
页码:31 / 38
页数:8
相关论文
共 29 条
  • [1] Deep Learning for Hate Speech Detection in Tweets
    Badjatiya, Pinkesh
    Gupta, Shashank
    Gupta, Manish
    Varma, Vasudeva
    [J]. WWW'17 COMPANION: PROCEEDINGS OF THE 26TH INTERNATIONAL CONFERENCE ON WORLD WIDE WEB, 2017, : 759 - 760
  • [2] Bahdanau D, 2016, Arxiv, DOI arXiv:1409.0473
  • [3] Building deep learning models for evidence classification from the open access biomedical literature
    Burns, Gully A.
    Li, Xiangci
    Peng, Nanyun
    [J]. DATABASE-THE JOURNAL OF BIOLOGICAL DATABASES AND CURATION, 2019,
  • [4] Chen Y., 2020, P 58 ANN M ASS COMPU, P7893, DOI 10.18653/v1/2020.acl-main.705
  • [5] Cohan Arman, 2018, P 2018 C N AM CHAPTE, V2, P615
  • [6] Devlin J., P NAACL HLT 2019 ASS, P4171
  • [7] dHoffschmidt M., 2019, P 42 INT ACM SIGIR C, P1133
  • [8] COVID-19 information retrieval with deep-learning based semantic search, question answering, and abstractive summarization
    Esteva, Andre
    Kale, Anuprit
    Paulus, Romain
    Hashimoto, Kazuma
    Yin, Wenpeng
    Radev, Dragomir
    Socher, Richard
    [J]. NPJ DIGITAL MEDICINE, 2021, 4 (01)
  • [9] Supervised Machine Learning Algorithms for Bioelectromagnetics: Prediction Models and Feature Selection Techniques Using Data from Weak Radiofrequency Radiation Effect on Human and Animals Cells
    Halgamuge, Malka N.
    [J]. INTERNATIONAL JOURNAL OF ENVIRONMENTAL RESEARCH AND PUBLIC HEALTH, 2020, 17 (12) : 1 - 27
  • [10] Long short-term memory
    Hochreiter, S
    Schmidhuber, J
    [J]. NEURAL COMPUTATION, 1997, 9 (08) : 1735 - 1780