Enhancing abstractive summarization of implicit datasets with contrastive attention

被引:1
|
作者
Kwon S. [1 ]
Lee Y. [2 ]
机构
[1] Department of Data Science, Seoul National University of Science and Technology, 232, Gongneung-ro, Nowon-gu, Seoul
[2] Department of Industrial Engineering, Seoul National University of Science and Technology, 232, Gongneung-ro, Nowon-gu, Seoul
基金
新加坡国家研究基金会;
关键词
Abstractive summarization; Contrastive attention; Implicit dataset; Text summarization;
D O I
10.1007/s00521-024-09864-y
中图分类号
学科分类号
摘要
It is important for abstractive summarization models to understand the important parts of the original document and create a natural summary accordingly. Recently, studies have been conducted to incorporate important parts of the original document during learning and have shown good performance. However, these studies are effective for explicit datasets but not implicit datasets which are relatively more abstract. This study addresses the challenge of summarizing implicit datasets, which have a lower deviation in the significance of important sentences compared to explicit datasets. A multi-task learning approach that reflects information about salient and incidental objects during the learning process was proposed. This was achieved by adding a contrastive objective to the fine-tuning process of the encoder-decoder language model. The salient and incidental parts were selected based on the ROUGE-L F1 score and their relationships were learned through triplet loss. The proposed method was evaluated using five benchmark summarization datasets, including two explicit and three implicit. The experimental results showed a greater improvement in implicit datasets, particularly for the highly abstractive XSum dataset, compared to the vanilla fine-tuning method in both the BART-base and T5-small models. © The Author(s), under exclusive licence to Springer-Verlag London Ltd., part of Springer Nature 2024.
引用
收藏
页码:15337 / 15351
页数:14
相关论文
共 50 条
  • [31] Learning Cluster Patterns for Abstractive Summarization
    Jo, Sung-Guk
    Park, Seung-Hyeok
    Kim, Jeong-Jae
    On, Byung-Won
    IEEE ACCESS, 2023, 11 : 146065 - 146075
  • [32] ABSUM: ABstractive SUMmarization of Lecture Videos
    Devi, M. S. Karthika
    Bhuvaneshwari, R.
    Baskaran, R.
    SMART TRENDS IN COMPUTING AND COMMUNICATIONS, VOL 3, SMARTCOM 2024, 2024, 947 : 237 - 248
  • [33] A Combined Extractive With Abstractive Model for Summarization
    Liu, Wenfeng
    Gao, Yaling
    Li, Jinming
    Yang, Yuzhen
    IEEE ACCESS, 2021, 9 : 43970 - 43980
  • [34] Abstractive vs. Extractive Summarization: An Experimental Review
    Giarelis, Nikolaos
    Mastrokostas, Charalampos
    Karacapilidis, Nikos
    APPLIED SCIENCES-BASEL, 2023, 13 (13):
  • [35] Natural Language Inference as an Evaluation Measure for Abstractive Summarization
    Bora-Kathariya, Rajeshree
    Haribhakta, Yashodhara
    2018 4TH INTERNATIONAL CONFERENCE FOR CONVERGENCE IN TECHNOLOGY (I2CT), 2018,
  • [36] Graph-based abstractive biomedical text summarization
    Givchi, Azadeh
    Ramezani, Reza
    Baraani-Dastjerdi, Ahmad
    JOURNAL OF BIOMEDICAL INFORMATICS, 2022, 132
  • [37] Reducing repetition in convolutional abstractive summarization
    Liu, Yizhu
    Chen, Xinyue
    Luo, Xusheng
    Zhu, Kenny Q.
    NATURAL LANGUAGE ENGINEERING, 2023, 29 (01) : 81 - 109
  • [38] Abstractive Summarization Model with Adaptive Sparsemax
    Guo, Shiqi
    Si, Yumeng
    Zhao, Jing
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, NLPCC 2022, PT I, 2022, 13551 : 810 - 821
  • [39] A Semantic Supervision Method for Abstractive Summarization
    Hu, Sunqiang
    Li, Xiaoyu
    Deng, Yu
    Peng, Yu
    Lin, Bin
    Yang, Shan
    CMC-COMPUTERS MATERIALS & CONTINUA, 2021, 69 (01): : 145 - 158
  • [40] ASoVS: Abstractive Summarization of Video Sequences
    Dilawari, Aniqa
    Khan, Muhammad Usman Ghani
    IEEE ACCESS, 2019, 7 : 29253 - 29263