Event detection by combining self-attention and CNN-BiGRU

被引:0
|
作者
Wang K. [1 ]
Wang M. [2 ]
Liu X. [1 ]
Tian G. [3 ]
Li C. [3 ]
Liu W. [2 ]
机构
[1] The 10th Research Institute of China Electronics Technology Group Corporation, Chengdu
[2] School of Telecommunications Engineering, Xidian University, Xi'an
[3] School of Computer Science and Technology, Xi'an University of Posts and Telecommunications, Xi'an
关键词
bidirectional gated recurrent unit; convolutional neural networks; event detection; information extraction; self-attention mechanism;
D O I
10.19665/j.issn1001-2400.2022.05.021
中图分类号
TN911 [通信理论];
学科分类号
081002 ;
摘要
Event detection methods based on convolutional neural networks and recurrent neural networks have been widely investigated. However, convolutional neural networks only consider local information within the convolution window, ignoring the context of words. Recurrent neural networks have the problem of vanishing gradient and short-term memory, and their variant gated recurrent units cannot get the features of each word. Therefore, in this paper, an event detection method based on self-attention and convolutional bidirectional gated recurrent units model is proposed, which takes both word vectors and position vectors as inputs. It can not only extract vocabulary level features with different granularities by convolutional neural network and sentence level features by bidirectional gated recurrent units, but also consider global information and pay attention to more important features for event detection by self-attention. The extracted lexical-level features and sentence-level features are combined as the joint features, and the candidate words are classified by the softmax classifier to complete the event detection task. Experimental results show that the F scores of trigger words recognition and classification reach 78. 9% and 76. 0% respectively on the ACE2005 English corpus, which are better than the results of benchmark methods. Furthermore, the model shows great convergence. It is shown that the proposed model based on self-attention and convolutional bidirectional gated recurrent units possesses good ability of text feature extraction and improves the performance of event detection. © 2022 Science Press. All rights reserved.
引用
收藏
页码:181 / 188
页数:7
相关论文
共 19 条
  • [1] WEI H, ZHOU A, ZHANG Y J., Biomedical Event Trigger Extraction Based on Multi-Layer Residual BiLSTM and Contextualized Word Representations, International Journal of Machine Learning and Cybernetics, 12, 18, pp. 1-13, (2021)
  • [2] XIANGW, WANG B., A Survey of Event Extraction From Text [J], IEEE Access, 7, pp. 173111-173137, (2019)
  • [3] WAN Qizhi, WAN Changxuan, HU Rong, Et al., Chinese Financial Event Extraction Based on Syntactic and Semantic Dependency Parsing [J], Chinese Journal of Computers, 44, 3, pp. 508-530, (2021)
  • [4] YU X Y, RONG W G, ZHOU D. Y, Et al., LSTM-Based End-to-End Framework for Biomedical Event Extraction [J], IEEE/ACM Transactions on Computational Biology and Bioinformatics, 17, 6, pp. 2029-2039, (2020)
  • [5] Ruifang HE, DUAN Shaoyang, Joint Chinese Event Extraction Based Multi-task Learning [J], Journal ol Software, 30, 4, pp. 1015-1030, (2019)
  • [6] SHEN Lanben, WU Zhihao, JI Yuze, Et al., Chinese Event Detection Method Combining Attention Mechanism and BiLSTM[J], Journal of Chinese Information Processing, 33, 9, pp. 79-87, (2019)
  • [7] ZHAN L Y, JIANG X P, LIU Q., Research on Chinese Event Extraction Method Based on HMM and Multi-stage Method, Journal of Physics: Conference Series, 1732, 1, pp. 1-4, (2021)
  • [8] AHN D., The Stage of Event Extraction, Proceedings of the Workshop on Annotating and Reasoning about Time and Events, pp. 1-8, (2006)
  • [9] LI Q, JI H, HUANG L., Joint Event Extraction via Structured Prediction with Global Features [C], Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics, pp. 73-82, (2013)
  • [10] CHEN Y B, XU L H, LIU K, Et al., Event Extraction via Dynamic Multi-Pooling Convolutional Neural Networks[C], Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing, pp. 167-176, (2015)