Sentence salience contrastive learning for abstractive text summarization

被引:6
作者
Huang, Ying [1 ,2 ]
Li, Zhixin [1 ,2 ]
Chen, Zhenbin [1 ,2 ]
Zhang, Canlong [1 ,2 ]
Ma, Huifang [3 ]
机构
[1] Guangxi Normal Univ, Key Lab Educ Blockchain & Intelligent Technol, Minist Educ, Guilin 541004, Peoples R China
[2] Guangxi Normal Univ, Guangxi Key Lab Multisource Informat Min & Secur, Guilin 541004, Peoples R China
[3] Northwest Normal Univ, Coll Comp Sci & Engn, Lanzhou 730070, Peoples R China
基金
中国国家自然科学基金;
关键词
Contrastive learning; Abstractive text summarization; Semantic similarity; Sentence salience; NETWORKS;
D O I
10.1016/j.neucom.2024.127808
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Text summarization aims to generate a short summary for a document while preserving salient information. Recently, contrastive learning has been extended from visual representation to summarization tasks. At present, the methods of contrastive learning summarization focus on modeling the global semantics of source documents, targets and candidate summaries to maximize their similarities. However, they ignore the influence of sentence semantics in the document. In this paper, we propose a sentence-level salience contrastive learning method to help the model capture salient information and denoise. The model expresses the sentence salience according to the semantic similarity between the summaries and sentences of the source document, and integrates the similarity distance into the contrastive loss in the form of soft weights. Therefore, our model maximize the similarity between summaries and salient information, while minimizing the similarity between summaries and potential noise. We have verified our method in three widely used datasets, CNN/Daily Mail, XSum and PubMed. The experimental results show that the proposed method can significantly improve the baseline performance and achieve competitive performance in the existing contrastive learning methods.
引用
收藏
页数:13
相关论文
共 50 条
  • [41] A Survey of the State-of-the-Art Models in Neural Abstractive Text Summarization
    Syed, Ayesha Ayub
    Gaol, Ford Lumban
    Matsuo, Tokuro
    IEEE ACCESS, 2021, 9 : 13248 - 13265
  • [42] Contrastive Learning Models for Sentence Representations
    Xu, Lingling
    Xie, Haoran
    Li, Zongxi
    Wang, Fu Lee
    Wang, Weiming
    Li, Qing
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2023, 14 (04)
  • [43] Entity Relations Based Pointer-Generator Network for Abstractive Text Summarization
    Huang, Tiancheng
    Lu, Guangquan
    Li, Zexin
    Song, Jiagang
    Wu, Lijuan
    ADVANCED DATA MINING AND APPLICATIONS, ADMA 2021, PT II, 2022, 13088 : 219 - 236
  • [44] Multi-layered attentional peephole convolutional LSTM for abstractive text summarization
    Rahman, Md Motiur
    Siddiqui, Fazlul Hasan
    ETRI JOURNAL, 2021, 43 (02) : 288 - 298
  • [45] Hindi Abstractive Text Summarization using Transliteration with Pre-trained Model
    Kumar, Jeetendra
    Shekhar, Shashi
    Gupta, Rashmi
    JOURNAL OF ELECTRICAL SYSTEMS, 2024, 20 (03) : 2089 - 2110
  • [46] Pointer-Generator Abstractive Text Summarization Model with Part of Speech Features
    Ren, Shuxia
    Zhang, Zheming
    PROCEEDINGS OF 2019 IEEE 10TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING AND SERVICE SCIENCE (ICSESS 2019), 2019, : 514 - 517
  • [47] Improving colloquial case legal judgment prediction via abstractive text summarization
    Hong, Yu-Xiang
    Chang, Chia-Hui
    COMPUTER LAW & SECURITY REVIEW, 2023, 51
  • [48] A Determinantal Point Process Based Novel Sampling Method of Abstractive Text Summarization
    Shen, Jianbin
    Xuan, Junyu
    Liang, Christy
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [49] See, hear, read: Leveraging multimodality with guided attention for abstractive text summarization
    Atri, Yash Kumar
    Pramanick, Shraman
    Goyal, Vikram
    Chakraborty, Tanmoy
    KNOWLEDGE-BASED SYSTEMS, 2021, 227
  • [50] Multilayer encoder and single-layer decoder for abstractive Arabic text summarization
    Suleiman, Dima
    Awajan, Arafat
    KNOWLEDGE-BASED SYSTEMS, 2022, 237