Improving abstractive summarization based on dynamic residual network with reinforce dependency

被引:11
|
作者
Liao, Weizhi [1 ]
Ma, Yaheng [1 ]
Yin, Yanchao [2 ]
Ye, Guanglei [1 ]
Zuo, Dongzhou [1 ]
机构
[1] Univ Elect Sci & Technol China, Chengdu, Peoples R China
[2] Kunming Univ Sci & Technol, Kunming, Yunnan, Peoples R China
基金
国家重点研发计划;
关键词
Abstractive summarization; Dynamic residual network; Reinforcement learning agent; Long-term dependencies; One-dimensional convolution; Sequence-to-sequence;
D O I
10.1016/j.neucom.2021.02.028
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The Seq2Seq abstract summarization model based on long short-term memory (LSTM) is very effective for short text summarization. However, LSTM is limited by long-term dependencies, which can potentially result in salient information loss when long text is processed by the Seq2Seq model based on LSTM. To overcome the long-term dependence limitation, an encoder-decoder model based on the dynamic residual network is proposed in this work. The model can dynamically select an optimal state from the state history to establish a connection with the current state to improve the LSTM long sequence dependencies according to the current decoding environment. Because the dynamic residual connections will result in long-term connection-dependent words, a new method based on reinforcement learning is proposed to simulate the dependence between words, which is then implemented into the training process of the model. This model is verified using the CNN/Daily Mail and New York Times datasets, and the experimental results show that the proposed model achieves significant improvements in capturing longterm dependencies compared with the traditional LSTM-based Seq2Seq abstractive summarization model.& nbsp; (c) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页码:228 / 237
页数:10
相关论文
共 48 条
  • [1] Improving Abstractive Summarization with Unsupervised Dynamic LoRA Mixtures
    Chernyshev, D. I.
    LOBACHEVSKII JOURNAL OF MATHEMATICS, 2024, 45 (07) : 2995 - 3006
  • [2] Improving Abstractive Summarization with Iterative Representation
    Li, Jinpeng
    Zhang, Chuang
    Chen, Xiaojun
    Cao, Yanan
    Jia, Ruipeng
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [3] Abstractive Text Summarization Based on Semantic Alignment Network
    Wu S.
    Huang D.
    Li J.
    Beijing Daxue Xuebao (Ziran Kexue Ban)/Acta Scientiarum Naturalium Universitatis Pekinensis, 2021, 57 (01): : 1 - 6
  • [4] Improving Pointer-Generator Network with Keywords Information for Chinese Abstractive Summarization
    Jiang, Xiaoping
    Hu, Po
    Hou, Liwei
    Wang, Xia
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, PT I, 2018, 11108 : 464 - 474
  • [5] Attention based Abstractive Summarization of Malayalam Document
    Nambiar, Sindhya K.
    Peter, David S.
    Idicula, Sumam Mary
    AI IN COMPUTATIONAL LINGUISTICS, 2021, 189 : 250 - 257
  • [6] Recurrent neural network for abstractive summarization of documents
    Bansal, Neha
    Sharma, Arun
    Singh, R. K.
    JOURNAL OF DISCRETE MATHEMATICAL SCIENCES & CRYPTOGRAPHY, 2020, 23 (01) : 65 - 72
  • [7] Warm-Starting for Improving the Novelty of Abstractive Summarization
    Alomari, Ayham
    Al-Shamayleh, Ahmad Sami
    Idris, Norisma
    Qalid Md Sabri, Aznul
    Alsmadi, Izzat
    Omary, Danah
    IEEE ACCESS, 2023, 11 : 112483 - 112501
  • [8] Improving Abstractive Dialogue Summarization Using Keyword Extraction
    Yoo, Chongjae
    Lee, Hwanhee
    APPLIED SCIENCES-BASEL, 2023, 13 (17):
  • [9] Extractive Elementary Discourse Units for Improving Abstractive Summarization
    Xiong, Ye
    Racharak, Teeradaj
    Minh Le Nguyen
    PROCEEDINGS OF THE 45TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '22), 2022, : 2675 - 2679
  • [10] Abstractive Multi-Document Summarization Based on Semantic Link Network
    Li, Wei
    Zhuge, Hai
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2021, 33 (01) : 43 - 54