Warm-Starting for Improving the Novelty of Abstractive Summarization

被引:0
|
作者
Alomari, Ayham [1 ,2 ]
Al-Shamayleh, Ahmad Sami
Idris, Norisma [3 ]
Qalid Md Sabri, Aznul [4 ]
Alsmadi, Izzat [5 ]
Omary, Danah [6 ]
机构
[1] Appl Sci Private Univ, Fac Informat Technol, Dept Comp Sci, Amman 11931, Jordan
[2] Middle East Univ, MEU Res Unit, Amman 11831, Jordan
[3] Al Ahliyya Amman Univ, Dept Data Sci & Artificial Intelligence, Amman 19328, Jordan
[4] Univ Malaya, Fac Comp Sci & Informat Technol, Dept Artificial Intelligence, Kuala Lumpur 50603, Malaysia
[5] Texas A&M Univ San Antonio, Dept Comp & Cybersecur, San Antonio, TX 78224 USA
[6] Univ North Texas, Dept Elect Engn, Denton, TX 76210 USA
关键词
Abstractive summarization; novelty; warm-started models; deep learning; metrics;
D O I
10.1109/ACCESS.2023.3322226
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The Abstractive summarization is distinguished by using novel phrases that are not found in the source text. However, most previous research ignores this feature in favour of enhancing syntactical similarity with the reference. To improve novelty aspects, we have used multiple warm-started models with varying encoder and decoder checkpoints and vocabulary. These models are then adapted to the paraphrasing task and the sampling decoding strategy to further boost the levels of novelty and quality. In addition, to avoid relying only on the syntactical similarity assessment, two additional abstractive summarization metrics are introduced: 1) NovScore: a new novelty metric that delivers a summary novelty score; and 2) NSSF: a new comprehensive metric that ensembles Novelty, Syntactic, Semantic, and Faithfulness features into a single score to simulate human assessment in providing a reliable evaluation. Finally, we compare our models to the state-of-the-art sequence-to-sequence models using the current and the proposed metrics. As a result, warm-starting, sampling, and paraphrasing improve novelty degrees by 2%, 5%, and 14%, respectively, while maintaining comparable scores on other metrics.
引用
收藏
页码:112483 / 112501
页数:19
相关论文
共 50 条
  • [21] ABSUM: ABstractive SUMmarization of Lecture Videos
    Devi, M. S. Karthika
    Bhuvaneshwari, R.
    Baskaran, R.
    SMART TRENDS IN COMPUTING AND COMMUNICATIONS, VOL 3, SMARTCOM 2024, 2024, 947 : 237 - 248
  • [22] Neural abstractive summarization fusing by global generative topics
    Gao, Yang
    Wang, Yang
    Liu, Luyang
    Guo, Yidi
    Huang, Heyan
    NEURAL COMPUTING & APPLICATIONS, 2020, 32 (09) : 5049 - 5058
  • [23] Abstractive vs. Extractive Summarization: An Experimental Review
    Giarelis, Nikolaos
    Mastrokostas, Charalampos
    Karacapilidis, Nikos
    APPLIED SCIENCES-BASEL, 2023, 13 (13):
  • [24] Reducing repetition in convolutional abstractive summarization
    Liu, Yizhu
    Chen, Xinyue
    Luo, Xusheng
    Zhu, Kenny Q.
    NATURAL LANGUAGE ENGINEERING, 2023, 29 (01) : 81 - 109
  • [25] Abstractive Summarization Model with Adaptive Sparsemax
    Guo, Shiqi
    Si, Yumeng
    Zhao, Jing
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, NLPCC 2022, PT I, 2022, 13551 : 810 - 821
  • [26] ASoVS: Abstractive Summarization of Video Sequences
    Dilawari, Aniqa
    Khan, Muhammad Usman Ghani
    IEEE ACCESS, 2019, 7 : 29253 - 29263
  • [27] A Study on Ontology based Abstractive Summarization
    Mohan, Jishma M.
    Sunitha, C.
    Ganesh, Amal
    Jaya, A.
    FOURTH INTERNATIONAL CONFERENCE ON RECENT TRENDS IN COMPUTER SCIENCE & ENGINEERING (ICRTCSE 2016), 2016, 87 : 32 - 37
  • [28] Abstractive Document Summarization via Bidirectional Decoder
    Wan, Xin
    Li, Chen
    Wang, Ruijia
    Xiao, Ding
    Shi, Chuan
    ADVANCED DATA MINING AND APPLICATIONS, ADMA 2018, 2018, 11323 : 364 - 377
  • [29] Variational Neural Decoder for Abstractive Text Summarization
    Zhao, Huan
    Cao, Jie
    Xu, Mingquan
    Lu, Jian
    COMPUTER SCIENCE AND INFORMATION SYSTEMS, 2020, 17 (02) : 537 - 552
  • [30] Abstractive Summarizers Become Emotional on News Summarization
    Ahuir, Vicent
    Gonzalez, Jose-Angel
    Hurtado, Lluis-F.
    Segarra, Encarna
    APPLIED SCIENCES-BASEL, 2024, 14 (02):