Improving abstractive summarization based on dynamic residual network with reinforce dependency

被引:11
|
作者
Liao, Weizhi [1 ]
Ma, Yaheng [1 ]
Yin, Yanchao [2 ]
Ye, Guanglei [1 ]
Zuo, Dongzhou [1 ]
机构
[1] Univ Elect Sci & Technol China, Chengdu, Peoples R China
[2] Kunming Univ Sci & Technol, Kunming, Yunnan, Peoples R China
基金
国家重点研发计划;
关键词
Abstractive summarization; Dynamic residual network; Reinforcement learning agent; Long-term dependencies; One-dimensional convolution; Sequence-to-sequence;
D O I
10.1016/j.neucom.2021.02.028
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The Seq2Seq abstract summarization model based on long short-term memory (LSTM) is very effective for short text summarization. However, LSTM is limited by long-term dependencies, which can potentially result in salient information loss when long text is processed by the Seq2Seq model based on LSTM. To overcome the long-term dependence limitation, an encoder-decoder model based on the dynamic residual network is proposed in this work. The model can dynamically select an optimal state from the state history to establish a connection with the current state to improve the LSTM long sequence dependencies according to the current decoding environment. Because the dynamic residual connections will result in long-term connection-dependent words, a new method based on reinforcement learning is proposed to simulate the dependence between words, which is then implemented into the training process of the model. This model is verified using the CNN/Daily Mail and New York Times datasets, and the experimental results show that the proposed model achieves significant improvements in capturing longterm dependencies compared with the traditional LSTM-based Seq2Seq abstractive summarization model.& nbsp; (c) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页码:228 / 237
页数:10
相关论文
共 48 条
  • [41] A hierarchical framework based on transformer technology to achieve factual consistent and non-redundant abstractive text summarization
    G. Swetha
    S. Phani Kumar
    Multimedia Tools and Applications, 2024, 83 : 47587 - 47608
  • [42] A hierarchical framework based on transformer technology to achieve factual consistent and non-redundant abstractive text summarization
    Swetha, G.
    Kumar, S. Phani
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 83 (16) : 47587 - 47608
  • [43] A Performance Analysis of Deep-Learning-Based Thai News Abstractive Summarization: Word Positions and Document Length
    Jumpathong, Sawittree
    Theeramunkong, Thanaruk
    Supnithi, Thepchai
    Okumura, Manabu
    2022 7TH INTERNATIONAL CONFERENCE ON BUSINESS AND INDUSTRIAL RESEARCH (ICBIR2022), 2022, : 279 - 284
  • [44] A data package for abstractive opinion summarization, title generation, and rating-based sentiment prediction for airline reviews
    Syed, Ayesha Ayub
    Gaol, Ford Lumban
    Boediman, Alfred
    Matsuo, Tokuro
    Budiharto, Widodo
    DATA IN BRIEF, 2023, 50
  • [45] T5-Based Model for Abstractive Summarization: A Semi-Supervised Learning Approach with Consistency Loss Functions
    Wang, Mingye
    Xie, Pan
    Du, Yao
    Hu, Xiaohui
    APPLIED SCIENCES-BASEL, 2023, 13 (12):
  • [46] Improving Text Summarization Using Feature Extraction Approach Based on Pointer-generator with Coverage
    Chen, Yongchao
    He, Xin
    Wang, Guanghui
    Yu, Junyang
    WEB INFORMATION SYSTEMS AND APPLICATIONS (WISA 2021), 2021, 12999 : 489 - 496
  • [47] Transformer-Based Abstractive Summarization for Reddit and Twitter: Single Posts vs. Comment Pools in Three Languages
    Blekanov, Ivan S.
    Tarasov, Nikita
    Bodrunova, Svetlana S.
    FUTURE INTERNET, 2022, 14 (03)
  • [48] Bearing Fault Diagnosis Based on One-Dimensional Convolution Network and Residual Training
    Li, Bin
    Zhang, CaiXia
    Liu, GuoWen
    PROCEEDINGS OF THE 38TH CHINESE CONTROL CONFERENCE (CCC), 2019, : 5018 - 5023