Warm-Starting for Improving the Novelty of Abstractive Summarization

被引:0
|
作者
Alomari, Ayham [1 ,2 ]
Al-Shamayleh, Ahmad Sami
Idris, Norisma [3 ]
Qalid Md Sabri, Aznul [4 ]
Alsmadi, Izzat [5 ]
Omary, Danah [6 ]
机构
[1] Appl Sci Private Univ, Fac Informat Technol, Dept Comp Sci, Amman 11931, Jordan
[2] Middle East Univ, MEU Res Unit, Amman 11831, Jordan
[3] Al Ahliyya Amman Univ, Dept Data Sci & Artificial Intelligence, Amman 19328, Jordan
[4] Univ Malaya, Fac Comp Sci & Informat Technol, Dept Artificial Intelligence, Kuala Lumpur 50603, Malaysia
[5] Texas A&M Univ San Antonio, Dept Comp & Cybersecur, San Antonio, TX 78224 USA
[6] Univ North Texas, Dept Elect Engn, Denton, TX 76210 USA
关键词
Abstractive summarization; novelty; warm-started models; deep learning; metrics;
D O I
10.1109/ACCESS.2023.3322226
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The Abstractive summarization is distinguished by using novel phrases that are not found in the source text. However, most previous research ignores this feature in favour of enhancing syntactical similarity with the reference. To improve novelty aspects, we have used multiple warm-started models with varying encoder and decoder checkpoints and vocabulary. These models are then adapted to the paraphrasing task and the sampling decoding strategy to further boost the levels of novelty and quality. In addition, to avoid relying only on the syntactical similarity assessment, two additional abstractive summarization metrics are introduced: 1) NovScore: a new novelty metric that delivers a summary novelty score; and 2) NSSF: a new comprehensive metric that ensembles Novelty, Syntactic, Semantic, and Faithfulness features into a single score to simulate human assessment in providing a reliable evaluation. Finally, we compare our models to the state-of-the-art sequence-to-sequence models using the current and the proposed metrics. As a result, warm-starting, sampling, and paraphrasing improve novelty degrees by 2%, 5%, and 14%, respectively, while maintaining comparable scores on other metrics.
引用
收藏
页码:112483 / 112501
页数:19
相关论文
共 50 条
  • [1] Improving Coverage and Novelty of Abstractive Text Summarization Using Transfer Learning and Divide and Conquer Approaches
    Alomari, Ayham
    Idris, Norisma
    Sabri, Aznul Qalid Md
    Alsmadi, Izzat
    MALAYSIAN JOURNAL OF COMPUTER SCIENCE, 2023, 36 (03)
  • [2] Improving Abstractive Summarization with Iterative Representation
    Li, Jinpeng
    Zhang, Chuang
    Chen, Xiaojun
    Cao, Yanan
    Jia, Ruipeng
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [3] Improving Abstractive Summarization with Unsupervised Dynamic LoRA Mixtures
    Chernyshev, D. I.
    LOBACHEVSKII JOURNAL OF MATHEMATICS, 2024, 45 (07) : 2995 - 3006
  • [4] Improving Abstractive Dialogue Summarization Using Keyword Extraction
    Yoo, Chongjae
    Lee, Hwanhee
    APPLIED SCIENCES-BASEL, 2023, 13 (17):
  • [5] Extractive Elementary Discourse Units for Improving Abstractive Summarization
    Xiong, Ye
    Racharak, Teeradaj
    Minh Le Nguyen
    PROCEEDINGS OF THE 45TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '22), 2022, : 2675 - 2679
  • [6] Improving Transformer with Sequential Context Representations for Abstractive Text Summarization
    Cai, Tian
    Shen, Mengjun
    Peng, Huailiang
    Jiang, Lei
    Dai, Qiong
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING (NLPCC 2019), PT I, 2019, 11838 : 512 - 524
  • [7] Abstractive summarization: An overview of the state of the art
    Gupta, Som
    Gupta, S. K.
    EXPERT SYSTEMS WITH APPLICATIONS, 2019, 121 : 49 - 65
  • [8] A Combined Extractive With Abstractive Model for Summarization
    Liu, Wenfeng
    Gao, Yaling
    Li, Jinming
    Yang, Yuzhen
    IEEE ACCESS, 2021, 9 : 43970 - 43980
  • [9] Improving named entity correctness of abstractive summarization by generative negative sampling
    Chen, Zheng
    Lin, Hongyu
    COMPUTER SPEECH AND LANGUAGE, 2023, 81
  • [10] Improving Abstractive Summarization by Training Masked Out-of-Vocabulary Words
    Lee, Tae-Seok
    Lee, Hyun-Young
    Kang, Seung-Shik
    JOURNAL OF INFORMATION PROCESSING SYSTEMS, 2022, 18 (03): : 344 - 358