Neural Abstractive Text Summarization with Sequence-to-Sequence Models

被引:68
|
作者
Shi, Tian [1 ]
Keneshloo, Yaser [1 ]
Ramakrishnan, Naren [1 ]
Reddy, Chandan K. [1 ]
机构
[1] Virginia Tech, 900 N Glebe Road, Arlington,VA,22203, United States
来源
关键词
Surveys - Modeling languages - Computational linguistics - Deep learning - Text processing;
D O I
10.1145/3419106
中图分类号
学科分类号
摘要
In the past few years, neural abstractive text summarization with sequence-to-sequence (seq2seq) models have gained a lot of popularity. Many interesting techniques have been proposed to improve seq2seq models, making them capable of handling different challenges, such as saliency, fluency and human readability, and generate high-quality summaries. Generally speaking, most of these techniques differ in one of these three categories: Network structure, parameter inference, and decoding/generation. There are also other concerns, such as efficiency and parallelism for training a model. In this article, we provide a comprehensive literature survey on different seq2seq models for abstractive text summarization from the viewpoint of network structures, training strategies, and summary generation algorithms. Several models were first proposed for language modeling and generation tasks, such as machine translation, and later applied to abstractive text summarization. Hence, we also provide a brief review of these models. As part of this survey, we also develop an open source library, namely, Neural Abstractive Text Summarizer (NATS) toolkit, for the abstractive text summarization. An extensive set of experiments have been conducted on the widely used CNN/Daily Mail dataset to examine the effectiveness of several different neural network components. Finally, we benchmark two models implemented in NATS on the two recently released datasets, namely, Newsroom and Bytecup. © 2021 ACM.
引用
收藏
相关论文
共 50 条
  • [21] Utilizing Character and Word Embeddings for Text Normalization with Sequence-to-Sequence Models
    Watson, Daniel
    Zalmout, Nasser
    Habash, Nizar
    2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), 2018, : 837 - 843
  • [22] A Detailed Evaluation of Neural Sequence-to-Sequence Models for In-domain and Cross-domain Text Simplification
    Stajner, Sanja
    Nisioi, Sergiu
    PROCEEDINGS OF THE ELEVENTH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION (LREC 2018), 2018, : 3026 - 3033
  • [23] Sparse Sequence-to-Sequence Models
    Peters, Ben
    Niculae, Vlad
    Martins, Andre F. T.
    57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 1504 - 1519
  • [24] A Survey of the State-of-the-Art Models in Neural Abstractive Text Summarization
    Syed, Ayesha Ayub
    Gaol, Ford Lumban
    Matsuo, Tokuro
    IEEE ACCESS, 2021, 9 : 13248 - 13265
  • [25] Variational Neural Decoder for Abstractive Text Summarization
    Zhao, Huan
    Cao, Jie
    Xu, Mingquan
    Lu, Jian
    COMPUTER SCIENCE AND INFORMATION SYSTEMS, 2020, 17 (02) : 537 - 552
  • [26] Turkish Data-to-Text Generation Using Sequence-to-Sequence Neural Networks
    Demir, Seniz
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2023, 22 (02)
  • [27] Rescoring Sequence-to-Sequence Models for Text Line Recognition with CTC-Prefixes
    Wick, Christoph
    Zollner, Jochen
    Gruning, Tobias
    DOCUMENT ANALYSIS SYSTEMS, DAS 2022, 2022, 13237 : 260 - 274
  • [28] Sequence-to-Sequence Contrastive Learning for Text Recognition
    Aberdam, Aviad
    Litman, Ron
    Tsiper, Shahar
    Anschel, Oron
    Slossberg, Ron
    Mazor, Shai
    Manmatha, R.
    Perona, Pietro
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 15297 - 15307
  • [29] Reinforcement Learning Models for Abstractive Text Summarization
    Buciumas, Sergiu
    PROCEEDINGS OF THE 2019 ANNUAL ACM SOUTHEAST CONFERENCE (ACMSE 2019), 2019, : 270 - 271
  • [30] Sequence-to-Sequence Neural Net Models for Grapheme-to-Phoneme Conversion
    Yao, Kaisheng
    Zweig, Geoffrey
    16TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2015), VOLS 1-5, 2015, : 3330 - 3334