Neural Abstractive Text Summarization with Sequence-to-Sequence Models

被引:68
|
作者
Shi, Tian [1 ]
Keneshloo, Yaser [1 ]
Ramakrishnan, Naren [1 ]
Reddy, Chandan K. [1 ]
机构
[1] Virginia Tech, 900 N Glebe Road, Arlington,VA,22203, United States
来源
关键词
Surveys - Modeling languages - Computational linguistics - Deep learning - Text processing;
D O I
10.1145/3419106
中图分类号
学科分类号
摘要
In the past few years, neural abstractive text summarization with sequence-to-sequence (seq2seq) models have gained a lot of popularity. Many interesting techniques have been proposed to improve seq2seq models, making them capable of handling different challenges, such as saliency, fluency and human readability, and generate high-quality summaries. Generally speaking, most of these techniques differ in one of these three categories: Network structure, parameter inference, and decoding/generation. There are also other concerns, such as efficiency and parallelism for training a model. In this article, we provide a comprehensive literature survey on different seq2seq models for abstractive text summarization from the viewpoint of network structures, training strategies, and summary generation algorithms. Several models were first proposed for language modeling and generation tasks, such as machine translation, and later applied to abstractive text summarization. Hence, we also provide a brief review of these models. As part of this survey, we also develop an open source library, namely, Neural Abstractive Text Summarizer (NATS) toolkit, for the abstractive text summarization. An extensive set of experiments have been conducted on the widely used CNN/Daily Mail dataset to examine the effectiveness of several different neural network components. Finally, we benchmark two models implemented in NATS on the two recently released datasets, namely, Newsroom and Bytecup. © 2021 ACM.
引用
收藏
相关论文
共 50 条
  • [1] Turkish abstractive text summarization using pretrained sequence-to-sequence models
    Baykara, Batuhan
    Gungor, Tunga
    NATURAL LANGUAGE ENGINEERING, 2023, 29 (05) : 1275 - 1304
  • [2] Towards Sequence-to-Sequence Neural Model for Croatian Abstractive Summarization
    Davidovic, Vlatka
    Ipsic, Sanda Martincic
    CENTRAL EUROPEAN CONFERENCE ON INFORMATION AND INTELLIGENT SYSTEMS, CECIIS, 2023, : 309 - 315
  • [3] Towards neural abstractive clinical trial text summarization with sequence to sequence models
    Cintas, Celia
    Ogallo, William
    Walcott, Aisha
    Remy, Sekou L.
    Akinwande, Victor
    Osebe, Samuel
    2019 IEEE INTERNATIONAL CONFERENCE ON HEALTHCARE INFORMATICS (ICHI), 2019, : 388 - 390
  • [4] A Reinforced Topic-Aware Convolutional Sequence-to-Sequence Model for Abstractive Text Summarization
    Wang, Li
    Yao, Junlin
    Tao, Yunzhe
    Zhong, Li
    Liu, Wei
    Du, Qiang
    PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2018, : 4453 - 4460
  • [5] Abstractive Text Summarization: Enhancing Sequence-to-Sequence Models Using Word Sense Disambiguation and Semantic Content Generalization
    Kouris, Panagiotis
    Alexandridis, Georgios
    Stafylopatis, Andreas
    COMPUTATIONAL LINGUISTICS, 2021, 47 (04) : 813 - 859
  • [6] Multi-View Sequence-to-Sequence Models with Conversational Structure for Abstractive Dialogue Summarization
    Chen, Jiaao
    Yang, Diyi
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 4106 - 4118
  • [7] Abstractive method of text summarization with sequence to sequence RNNs
    Masum, Abu Kaisar Mohammad
    Abujar, Sheikh
    Talukder, Md Ashraful Islam
    Rabby, A. K. M. Shahariar Azad
    Hossain, Syed Akhter
    2019 10TH INTERNATIONAL CONFERENCE ON COMPUTING, COMMUNICATION AND NETWORKING TECHNOLOGIES (ICCCNT), 2019,
  • [8] Bengali abstractive text summarization using sequence to sequence RNNs
    Talukder, Md Ashraful Islam
    Abujar, Sheikh
    Masum, Abu Kaisar Mohammad
    Faisal, Fahad
    Hossain, Syed Akhter
    2019 10TH INTERNATIONAL CONFERENCE ON COMPUTING, COMMUNICATION AND NETWORKING TECHNOLOGIES (ICCCNT), 2019,
  • [9] Sequence-to-Sequence Models for Automated Text Simplification
    Botarleanu, Robert-Mihai
    Dascalu, Mihai
    Crossley, Scott Andrew
    McNamara, Danielle S.
    ARTIFICIAL INTELLIGENCE IN EDUCATION (AIED 2020), PT II, 2020, 12164 : 31 - 36
  • [10] BART-IT: An Efficient Sequence-to-Sequence Model for Italian Text Summarization
    La Quatra, Moreno
    Cagliero, Luca
    FUTURE INTERNET, 2023, 15 (01)