Enhancing abstractive summarization of implicit datasets with contrastive attention

被引:1
|
作者
Kwon S. [1 ]
Lee Y. [2 ]
机构
[1] Department of Data Science, Seoul National University of Science and Technology, 232, Gongneung-ro, Nowon-gu, Seoul
[2] Department of Industrial Engineering, Seoul National University of Science and Technology, 232, Gongneung-ro, Nowon-gu, Seoul
基金
新加坡国家研究基金会;
关键词
Abstractive summarization; Contrastive attention; Implicit dataset; Text summarization;
D O I
10.1007/s00521-024-09864-y
中图分类号
学科分类号
摘要
It is important for abstractive summarization models to understand the important parts of the original document and create a natural summary accordingly. Recently, studies have been conducted to incorporate important parts of the original document during learning and have shown good performance. However, these studies are effective for explicit datasets but not implicit datasets which are relatively more abstract. This study addresses the challenge of summarizing implicit datasets, which have a lower deviation in the significance of important sentences compared to explicit datasets. A multi-task learning approach that reflects information about salient and incidental objects during the learning process was proposed. This was achieved by adding a contrastive objective to the fine-tuning process of the encoder-decoder language model. The salient and incidental parts were selected based on the ROUGE-L F1 score and their relationships were learned through triplet loss. The proposed method was evaluated using five benchmark summarization datasets, including two explicit and three implicit. The experimental results showed a greater improvement in implicit datasets, particularly for the highly abstractive XSum dataset, compared to the vanilla fine-tuning method in both the BART-base and T5-small models. © The Author(s), under exclusive licence to Springer-Verlag London Ltd., part of Springer Nature 2024.
引用
收藏
页码:15337 / 15351
页数:14
相关论文
共 50 条
  • [1] Attention based Abstractive Summarization of Malayalam Document
    Nambiar, Sindhya K.
    Peter, David S.
    Idicula, Sumam Mary
    AI IN COMPUTATIONAL LINGUISTICS, 2021, 189 : 250 - 257
  • [2] Enhancing Abstractive Summarization with Pointer Generator Networks and Coverage Mechanisms in NLP
    Yarlagadda, Madhulika
    Nadendla, Hanumantha Rao
    IRANIAN JOURNAL OF SCIENCE AND TECHNOLOGY-TRANSACTIONS OF ELECTRICAL ENGINEERING, 2024, 48 (03) : 1301 - 1311
  • [3] Introducing bidirectional attention for autoregressive models in abstractive summarization
    Zhao, Jianfei
    Sun, Xin
    Feng, Chong
    INFORMATION SCIENCES, 2025, 689
  • [4] Gated Graph Neural Attention Networks for abstractive summarization
    Liang, Zeyu
    Du, Junping
    Shao, Yingxia
    Ji, Houye
    NEUROCOMPUTING, 2021, 431 : 128 - 136
  • [5] A Semantic Similarity Distance-Aware Contrastive Learning for Abstractive Summarization
    Huang, Ying
    Li, Zhixin
    PRICAI 2023: TRENDS IN ARTIFICIAL INTELLIGENCE, PT I, 2024, 14325 : 173 - 185
  • [6] Abstractive Summarization of Text Document in Malayalam Language: Enhancing Attention Model Using POS Tagging Feature
    Nambiar, Sindhya K.
    Peter, David S.
    Idicula, Sumam Mary
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2023, 22 (02)
  • [7] Neural attention model with keyword memory for abstractive document summarization
    Choi, YunSeok
    Kim, Dahae
    Lee, Jee-Hyong
    CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2020, 32 (18)
  • [8] Incorporating word attention with convolutional neural networks for abstractive summarization
    Yuan, Chengzhe
    Bao, Zhifeng
    Sanderson, Mark
    Tang, Yong
    WORLD WIDE WEB-INTERNET AND WEB INFORMATION SYSTEMS, 2020, 23 (01): : 267 - 287
  • [9] A novel abstractive summarization model based on topic-aware and contrastive learning
    Tang, Huanling
    Li, Ruiquan
    Duan, Wenhao
    Dou, Quansheng
    Lu, Mingyu
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024, 15 (12) : 5563 - 5577
  • [10] Incorporating word attention with convolutional neural networks for abstractive summarization
    Chengzhe Yuan
    Zhifeng Bao
    Mark Sanderson
    Yong Tang
    World Wide Web, 2020, 23 : 267 - 287