Abstractive text summarization: State of the art, challenges, and improvements

被引:4
作者
Shakil, Hassan [1 ]
Farooq, Ahmad [2 ]
Kalita, Jugal [1 ]
机构
[1] Univ Colorado, Dept Comp Sci, Colorado Springs, CO 80918 USA
[2] Univ Arkansas, Dept Elect & Comp Engn, Little Rock, AR 72204 USA
基金
美国国家科学基金会;
关键词
Automatic summarization; Abstractive summarization; Extractive summarization; Knowledge representation; Text generation; KNOWLEDGE;
D O I
10.1016/j.neucom.2024.128255
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Specifically focusing on the landscape of abstractive text summarization, as opposed to extractive techniques, this survey presents a comprehensive overview, delving into state-of-the-art techniques, prevailing challenges, and prospective research directions. We categorize the techniques into traditional sequenceto-sequence models, pre-trained large language models, reinforcement learning, hierarchical methods, and multi-modal summarization. Unlike prior works that did not examine complexities, scalability and comparisons of techniques in detail, this review takes a comprehensive approach encompassing state-of-the-art methods, challenges, solutions, comparisons, limitations and charts out future improvements - providing researchers an extensive overview to advance abstractive summarization research. We provide vital comparison tables across techniques categorized - offering insights into model complexity, scalability and appropriate applications. The paper highlights challenges such as inadequate meaning representation, factual consistency, controllable text summarization, cross-lingual summarization, and evaluation metrics, among others. Solutions leveraging knowledge incorporation and other innovative strategies are proposed to address these challenges. The paper concludes by highlighting emerging research areas like factual inconsistency, domain-specific, cross-lingual, multilingual, and long-document summarization, as well as handling noisy data. Our objective is to provide researchers and practitioners with a structured overview of the domain, enabling them to better understand the current landscape and identify potential areas for further research and improvement.
引用
收藏
页数:28
相关论文
共 50 条
  • [41] Abstractive Summarization Model with Adaptive Sparsemax
    Guo, Shiqi
    Si, Yumeng
    Zhao, Jing
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, NLPCC 2022, PT I, 2022, 13551 : 810 - 821
  • [42] Improving Coverage and Novelty of Abstractive Text Summarization Using Transfer Learning and Divide and Conquer Approaches
    Alomari, Ayham
    Idris, Norisma
    Sabri, Aznul Qalid Md
    Alsmadi, Izzat
    MALAYSIAN JOURNAL OF COMPUTER SCIENCE, 2023, 36 (03)
  • [43] Enhancing N-Gram Based Metrics with Semantics for Better Evaluation of Abstractive Text Summarization
    Jia-Wei He
    Wen-Jun Jiang
    Guo-Bang Chen
    Yu-Quan Le
    Xiao-Fei Ding
    Journal of Computer Science and Technology, 2022, 37 : 1118 - 1133
  • [44] Enhancing N-Gram Based Metrics with Semantics for Better Evaluation of Abstractive Text Summarization
    He, Jia-Wei
    Jiang, Wen-Jun
    Chen, Guo-Bang
    Le, Yu-Quan
    Ding, Xiao-Fei
    JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY, 2022, 37 (05) : 1118 - 1133
  • [45] ASoVS: Abstractive Summarization of Video Sequences
    Dilawari, Aniqa
    Khan, Muhammad Usman Ghani
    IEEE ACCESS, 2019, 7 : 29253 - 29263
  • [46] A Study on Ontology based Abstractive Summarization
    Mohan, Jishma M.
    Sunitha, C.
    Ganesh, Amal
    Jaya, A.
    FOURTH INTERNATIONAL CONFERENCE ON RECENT TRENDS IN COMPUTER SCIENCE & ENGINEERING (ICRTCSE 2016), 2016, 87 : 32 - 37
  • [47] Improving Abstractive Summarization with Iterative Representation
    Li, Jinpeng
    Zhang, Chuang
    Chen, Xiaojun
    Cao, Yanan
    Jia, Ruipeng
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [48] Abstractive Summarization Improved by WordNet-Based Extractive Sentences
    Xie, Niantao
    Li, Sujian
    Ren, Huiling
    Zhai, Qibin
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, PT I, 2018, 11108 : 404 - 415
  • [49] Abstractive Summarization of Text Document in Malayalam Language: Enhancing Attention Model Using POS Tagging Feature
    Nambiar, Sindhya K.
    Peter, David S.
    Idicula, Sumam Mary
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2023, 22 (02)
  • [50] Attention based Abstractive Summarization of Malayalam Document
    Nambiar, Sindhya K.
    Peter, David S.
    Idicula, Sumam Mary
    AI IN COMPUTATIONAL LINGUISTICS, 2021, 189 : 250 - 257