Improving abstractive summarization based on dynamic residual network with reinforce dependency

被引:11
|
作者
Liao, Weizhi [1 ]
Ma, Yaheng [1 ]
Yin, Yanchao [2 ]
Ye, Guanglei [1 ]
Zuo, Dongzhou [1 ]
机构
[1] Univ Elect Sci & Technol China, Chengdu, Peoples R China
[2] Kunming Univ Sci & Technol, Kunming, Yunnan, Peoples R China
基金
国家重点研发计划;
关键词
Abstractive summarization; Dynamic residual network; Reinforcement learning agent; Long-term dependencies; One-dimensional convolution; Sequence-to-sequence;
D O I
10.1016/j.neucom.2021.02.028
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The Seq2Seq abstract summarization model based on long short-term memory (LSTM) is very effective for short text summarization. However, LSTM is limited by long-term dependencies, which can potentially result in salient information loss when long text is processed by the Seq2Seq model based on LSTM. To overcome the long-term dependence limitation, an encoder-decoder model based on the dynamic residual network is proposed in this work. The model can dynamically select an optimal state from the state history to establish a connection with the current state to improve the LSTM long sequence dependencies according to the current decoding environment. Because the dynamic residual connections will result in long-term connection-dependent words, a new method based on reinforcement learning is proposed to simulate the dependence between words, which is then implemented into the training process of the model. This model is verified using the CNN/Daily Mail and New York Times datasets, and the experimental results show that the proposed model achieves significant improvements in capturing longterm dependencies compared with the traditional LSTM-based Seq2Seq abstractive summarization model.& nbsp; (c) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页码:228 / 237
页数:10
相关论文
共 48 条
  • [21] Abstractive Summarization Based on Fine-Grained Interpretable Matrix
    Wang H.
    Gao Y.
    Feng J.
    Hu M.
    Wang H.
    Bai Y.
    Beijing Daxue Xuebao (Ziran Kexue Ban)/Acta Scientiarum Naturalium Universitatis Pekinensis, 2021, 57 (01): : 23 - 30
  • [22] ASM: Augmentation-based Semantic Mechanism on Abstractive Summarization
    Ren, Weidong
    Zhou, Hao
    Liu, Gongshen
    Huan, Fei
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [23] HITS-based attentional neural model for abstractive summarization
    Cai, Xiaoyan
    Shi, Kaile
    Jiang, Yuehan
    Yang, Libin
    Liu, Sen
    KNOWLEDGE-BASED SYSTEMS, 2021, 222
  • [24] A novel semantic-enhanced generative adversarial network for abstractive text summarization
    Tham Vo
    Soft Computing, 2023, 27 : 6267 - 6280
  • [25] A novel semantic-enhanced generative adversarial network for abstractive text summarization
    Vo, Tham
    SOFT COMPUTING, 2023, 27 (10) : 6267 - 6280
  • [26] TASP : Topic-based abstractive summarization of Facebook text posts
    Benedetto, Irene
    La Quatra, Moreno
    Cagliero, Luca
    Vassio, Luca
    Trevisan, Martino
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 255
  • [27] Deep Learning-Based Abstractive Summarization for Brazilian Portuguese Texts
    Palola, Pedro H.
    de Rose, Gustavo H.
    Papa, Joao P.
    INTELLIGENT SYSTEMS, PT II, 2022, 13654 : 479 - 493
  • [28] Multi-document Abstractive Summarization Based on Predicate Argument Structure
    Alshaina, S.
    John, Ansamma
    Nath, Aneesh G.
    2017 IEEE INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING, INFORMATICS, COMMUNICATION AND ENERGY SYSTEMS (SPICES), 2017,
  • [29] Sentence Pair Embeddings Based Evaluation Metric for Abstractive and Extractive Summarization
    Akula, Ramya
    Garibay, Ivan
    LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2022, : 6009 - 6017
  • [30] Improving Coverage and Novelty of Abstractive Text Summarization Using Transfer Learning and Divide and Conquer Approaches
    Alomari, Ayham
    Idris, Norisma
    Sabri, Aznul Qalid Md
    Alsmadi, Izzat
    MALAYSIAN JOURNAL OF COMPUTER SCIENCE, 2023, 36 (03)