Warm-Starting for Improving the Novelty of Abstractive Summarization

被引:0
|
作者
Alomari, Ayham [1 ,2 ]
Al-Shamayleh, Ahmad Sami
Idris, Norisma [3 ]
Qalid Md Sabri, Aznul [4 ]
Alsmadi, Izzat [5 ]
Omary, Danah [6 ]
机构
[1] Appl Sci Private Univ, Fac Informat Technol, Dept Comp Sci, Amman 11931, Jordan
[2] Middle East Univ, MEU Res Unit, Amman 11831, Jordan
[3] Al Ahliyya Amman Univ, Dept Data Sci & Artificial Intelligence, Amman 19328, Jordan
[4] Univ Malaya, Fac Comp Sci & Informat Technol, Dept Artificial Intelligence, Kuala Lumpur 50603, Malaysia
[5] Texas A&M Univ San Antonio, Dept Comp & Cybersecur, San Antonio, TX 78224 USA
[6] Univ North Texas, Dept Elect Engn, Denton, TX 76210 USA
关键词
Abstractive summarization; novelty; warm-started models; deep learning; metrics;
D O I
10.1109/ACCESS.2023.3322226
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The Abstractive summarization is distinguished by using novel phrases that are not found in the source text. However, most previous research ignores this feature in favour of enhancing syntactical similarity with the reference. To improve novelty aspects, we have used multiple warm-started models with varying encoder and decoder checkpoints and vocabulary. These models are then adapted to the paraphrasing task and the sampling decoding strategy to further boost the levels of novelty and quality. In addition, to avoid relying only on the syntactical similarity assessment, two additional abstractive summarization metrics are introduced: 1) NovScore: a new novelty metric that delivers a summary novelty score; and 2) NSSF: a new comprehensive metric that ensembles Novelty, Syntactic, Semantic, and Faithfulness features into a single score to simulate human assessment in providing a reliable evaluation. Finally, we compare our models to the state-of-the-art sequence-to-sequence models using the current and the proposed metrics. As a result, warm-starting, sampling, and paraphrasing improve novelty degrees by 2%, 5%, and 14%, respectively, while maintaining comparable scores on other metrics.
引用
收藏
页码:112483 / 112501
页数:19
相关论文
共 50 条
  • [41] IWM-LSTM encoder for abstractive text summarization
    Gangundi R.
    Sridhar R.
    Multimedia Tools and Applications, 2025, 84 (09) : 5883 - 5904
  • [42] Enhancing abstractive summarization of implicit datasets with contrastive attention
    Kwon S.
    Lee Y.
    Neural Computing and Applications, 2024, 36 (25) : 15337 - 15351
  • [43] Topic Attentional Neural Network for Abstractive Document Summarization
    Liu, Hao
    Zheng, Hai-Tao
    Wang, Wei
    ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2019, PT II, 2019, 11440 : 70 - 81
  • [44] Abstractive Text Summarization Based on Semantic Alignment Network
    Wu S.
    Huang D.
    Li J.
    Beijing Daxue Xuebao (Ziran Kexue Ban)/Acta Scientiarum Naturalium Universitatis Pekinensis, 2021, 57 (01): : 1 - 6
  • [45] Abstractive text summarization using deep learning with a new Turkish summarization benchmark dataset
    Ertam, Fatih
    Aydin, Galip
    CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2022, 34 (09)
  • [46] Introducing bidirectional attention for autoregressive models in abstractive summarization
    Zhao, Jianfei
    Sun, Xin
    Feng, Chong
    INFORMATION SCIENCES, 2025, 689
  • [47] A Novel Framework for Semantic Oriented Abstractive Text Summarization
    Moratanch, N.
    Chitrakala, S.
    JOURNAL OF WEB ENGINEERING, 2018, 17 (08): : 675 - 716
  • [48] Gated Graph Neural Attention Networks for abstractive summarization
    Liang, Zeyu
    Du, Junping
    Shao, Yingxia
    Ji, Houye
    NEUROCOMPUTING, 2021, 431 : 128 - 136
  • [49] Deep Learning Based Abstractive Turkish News Summarization
    Karakoc, Enise
    Yilmaz, Burcu
    2019 27TH SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2019,
  • [50] CATS: Customizable Abstractive Topic-based Summarization
    Bahrainian, Seyed Ali
    Zerveas, George
    Crestani, Fabio
    Eickhoff, Carsten
    ACM TRANSACTIONS ON INFORMATION SYSTEMS, 2022, 40 (01)