AHNN: An Attention-Based Hybrid Neural Network for Sentence Modeling

被引:1
|
作者
Zhang, Xiaomin [1 ]
Huang, Li [1 ]
Qu, Hong [1 ]
机构
[1] Univ Elect Sci & Technol China, Sch Comp Sci & Engn, Chengdu 610054, Sichuan, Peoples R China
基金
美国国家科学基金会;
关键词
Nature Language Processing (NLP); Sentence modeling; News Headline Categorization; Convolutional neural networks; Recurrent neural networks;
D O I
10.1007/978-3-319-73618-1_63
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks (DNNs) are powerful models that achieved excellent performance on many fields, especially in Nature Language Processing (NLP). Convolutional neural networks (CNN) and Recurrent neural networks (RNN) are two mainstream architectures of DNNs, are wildly explored to handle NLP tasks. However, those two type models adopt totally different ways to work. CNN is supposed to be good at capturing local features while RNN is considered to be able to summarize global information. In this paper, we combine the strengths of both architectures and propose a hybird model AHNN: Attention-based hybrid Neural Network, and use it in sentence modeling study. The AHNN utilizes attention based bidirectional dynamic lstm to obtain a better representation of global sentence information, then uses a parallel convolutional layer which has three different size filters and a max pooling layer to obtain significant local information. Finally, the two results are used together to feed into an expert layer to obtain results. Experiments show that the proposed architecture AHNN is able to summarize the context of the sentence and capture significant local features of sentence which is important for sentence modeling. We evaluate the proposed architecture AHNN on NLPCC News Headline Categorization test set and achieve 0.8098 test accuracy, it is a competitive performance compare with other teams in this task.
引用
收藏
页码:731 / 740
页数:10
相关论文
共 50 条
  • [1] A Sentence Summarizer using Recurrent Neural Network and Attention-Based Encoder
    Kuremoto, Takashi
    Tsuruda, Takuji
    Mabu, Shingo
    Obayashi, Masanao
    PROCEEDINGS OF THE 2017 INTERNATIONAL CONFERENCE ON APPLIED MATHEMATICS, MODELING AND SIMULATION (AMMS 2017), 2017, 153 : 245 - 248
  • [2] Attention-Based SeriesNet: An Attention-Based Hybrid Neural Network Model for Conditional Time Series Forecasting
    Cheng, Yepeng
    Liu, Zuren
    Morimoto, Yasuhiko
    INFORMATION, 2020, 11 (06)
  • [3] Attention-based Convolutional Neural Networks for Sentence Classification
    Zhao, Zhiwei
    Wu, Youzheng
    17TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2016), VOLS 1-5: UNDERSTANDING SPEECH PROCESSING IN HUMANS AND MACHINES, 2016, : 705 - 709
  • [4] Cascade Dynamics Modeling with Attention-based Recurrent Neural Network
    Wang, Yongqing
    Shen, Huawei
    Liu, Shenghua
    Gao, Jinhua
    Cheng, Xueqi
    PROCEEDINGS OF THE TWENTY-SIXTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 2985 - 2991
  • [5] A feature-level attention-based deep neural network model for sentence embedding
    Bouraoui A.
    Jamoussi S.
    Hamadou A.B.
    International Journal of Intelligent Systems Technologies and Applications, 2022, 20 (05) : 414 - 435
  • [6] AHRNN: Attention-Based Hybrid Robust Neural Network for emotion recognition
    Xu, Ke
    Liu, Bin
    Tao, Jianhua
    Lv, Zhao
    Fan, Cunhang
    Song, Leichao
    COGNITIVE COMPUTATION AND SYSTEMS, 2022, 4 (01) : 85 - 95
  • [7] Attention-based LSTM (AttLSTM) neural network for Seismic Response Modeling of Bridges
    Liao, Yuchen
    Lin, Rong
    Zhang, Ruiyang
    Wu, Gang
    COMPUTERS & STRUCTURES, 2023, 275
  • [8] Contaminant Transport Modeling and Source Attribution With Attention-Based Graph Neural Network
    Pang, Min
    Du, Erhu
    Zheng, Chunmiao
    WATER RESOURCES RESEARCH, 2024, 60 (06)
  • [9] Attention-Based Memory Network for Sentence-Level Question Answering
    Liu, Pei
    Zhang, Chunhong
    Zhang, Weiming
    Zhan, Zhiqiang
    Zhuang, Benhui
    SOCIAL MEDIA PROCESSING, SMP 2017, 2017, 774 : 104 - 115
  • [10] INTERPRETABLE GLUCOSE-LEVEL PREDICTION WITH HYBRID ATTENTION-BASED DIFFUSION NEURAL NETWORK
    Rahim, M.
    Blampey, Q.
    DIABETES TECHNOLOGY & THERAPEUTICS, 2023, 25 : A91 - A91