Biomedical word sense disambiguation with bidirectional long short-term memory and attention-based neural networks

被引:13
|
作者
Zhang, Canlin [1 ]
Bis, Daniel [2 ]
Liu, Xiuwen [2 ]
He, Zhe [3 ]
机构
[1] Florida State Univ, Dept Math, Tallahassee, FL 32306 USA
[2] Florida State Univ, Dept Comp Sci, Tallahassee, FL 32306 USA
[3] Florida State Univ, Sch Informat, Tallahassee, FL 32306 USA
关键词
Word sense disambiguation; LSTM; Self-attention; Biomedical;
D O I
10.1186/s12859-019-3079-8
中图分类号
Q5 [生物化学];
学科分类号
071010 ; 081704 ;
摘要
Background In recent years, deep learning methods have been applied to many natural language processing tasks to achieve state-of-the-art performance. However, in the biomedical domain, they have not out-performed supervised word sense disambiguation (WSD) methods based on support vector machines or random forests, possibly due to inherent similarities of medical word senses. Results In this paper, we propose two deep-learning-based models for supervised WSD: a model based on bi-directional long short-term memory (BiLSTM) network, and an attention model based on self-attention architecture. Our result shows that the BiLSTM neural network model with a suitable upper layer structure performs even better than the existing state-of-the-art models on the MSH WSD dataset, while our attention model was 3 or 4 times faster than our BiLSTM model with good accuracy. In addition, we trained "universal" models in order to disambiguate all ambiguous words together. That is, we concatenate the embedding of the target ambiguous word to the max-pooled vector in the universal models, acting as a "hint". The result shows that our universal BiLSTM neural network model yielded about 90 percent accuracy. Conclusion Deep contextual models based on sequential information processing methods are able to capture the relative contextual information from pre-trained input word embeddings, in order to provide state-of-the-art results for supervised biomedical WSD tasks.
引用
收藏
页数:15
相关论文
共 50 条
  • [21] Attention-Based Joint Learning for Intent Detection and Slot Filling Using Bidirectional Long Short-Term Memory and Convolutional Neural Networks (AJLISBC)
    Muhammad, Yusuf Idris
    Salim, Naomie
    Huspi, Sharin Hazlin
    Zainal, Anazida
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2024, 15 (08) : 915 - 922
  • [22] Attention-based Bidirectional Long Short-Term Memory Networks for Relation Classification Using Knowledge Distillation from BERT
    Wang, Zihan
    Yang, Bo
    2020 IEEE INTL CONF ON DEPENDABLE, AUTONOMIC AND SECURE COMPUTING, INTL CONF ON PERVASIVE INTELLIGENCE AND COMPUTING, INTL CONF ON CLOUD AND BIG DATA COMPUTING, INTL CONF ON CYBER SCIENCE AND TECHNOLOGY CONGRESS (DASC/PICOM/CBDCOM/CYBERSCITECH), 2020, : 562 - 568
  • [23] Short-Term Traffic Congestion Forecasting Using Attention-Based Long Short-Term Memory Recurrent Neural Network
    Zhang, Tianlin
    Liu, Ying
    Cui, Zhenyu
    Leng, Jiaxu
    Xie, Weihong
    Zhang, Liang
    COMPUTATIONAL SCIENCE - ICCS 2019, PT III, 2019, 11538 : 304 - 314
  • [24] Attention-Based Convolution Skip Bidirectional Long Short-Term Memory Network for Speech Emotion Recognition
    Zhang, Huiyun
    Huang, Heming
    Han, Henry
    IEEE ACCESS, 2021, 9 : 5332 - 5342
  • [25] Twitter Bot Detection Using Bidirectional Long Short-term Memory Neural Networks and Word Embeddings
    Wei, Feng
    Uyen Trang Nguyen
    2019 FIRST IEEE INTERNATIONAL CONFERENCE ON TRUST, PRIVACY AND SECURITY IN INTELLIGENT SYSTEMS AND APPLICATIONS (TPS-ISA 2019), 2019, : 101 - 109
  • [26] Supervised Attention-Based Bidirectional Long Short-Term Memory Network for Nonlinear Dynamic Soft Sensor Application
    Yang, Zeyu
    Jia, Ruining
    Wang, Peiliang
    Yao, Le
    Shen, Bingbing
    ACS OMEGA, 2023,
  • [27] Sarcasm Detection Using Soft Attention-Based Bidirectional Long Short-Term Memory Model With Convolution Network
    Le Hoang Son
    Kumar, Akshi
    Sangwan, Saurabh Raj
    Arora, Anshika
    Nayyar, Anand
    Abdel-Basset, Mohamed
    IEEE ACCESS, 2019, 7 : 23319 - 23328
  • [28] Arabic Word Segmentation With Long Short-Term Memory Neural Networks and Word Embedding
    Almuhareb, Abdulrahman
    Alsanie, Waleed
    Al-Thubaity, Abdulmohsen
    IEEE ACCESS, 2019, 7 : 12879 - 12887
  • [29] Attention-based convolutional neural network and long short-term memory for short-term detection of mood disorders based on elicited speech responses
    Huang, Kun-Yi
    Wu, Chung-Hsien
    Su, Ming-Hsiang
    PATTERN RECOGNITION, 2019, 88 : 668 - 678
  • [30] Attention-based recurrent neural networks for accurate short-term and long-term dissolved oxygen prediction
    Liu, Yeqi
    Zhang, Qian
    Song, Lihua
    Chen, Yingyi
    COMPUTERS AND ELECTRONICS IN AGRICULTURE, 2019, 165