DiffuSum: Generation Enhanced Extractive Summarization with Diffusion

被引:0
作者
Zhang, Haopeng [1 ]
Liu, Xiao [1 ]
Zhang, Jiawei [1 ]
机构
[1] Univ Calif Davis, IFM Lab, Dept Comp Sci, Davis, CA 95616 USA
来源
FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023) | 2023年
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Extractive summarization aims to form a summary by directly extracting sentences from the source document. Existing works mostly formulate it as a sequence labeling problem by making individual sentence label predictions. This paper proposes DiffuSum, a novel paradigm for extractive summarization, by directly generating the desired summary sentence representations with diffusion models and extracting sentences based on sentence representation matching. In addition, DiffuSum jointly optimizes a contrastive sentence encoder with a matching loss for sentence representation alignment and a multi-class contrastive loss for representation diversity. Experimental results show that DiffuSum achieves the new state-of-the-art extractive results on CNN/DailyMail with ROUGE scores of 44.83/22.56/40.56. Experiments on the other two datasets with different summary lengths also demonstrate the effectiveness of DiffuSum. The strong performance of our framework shows the great potential of adapting generative models for extractive summarization. To encourage more following work in the future, we have released our codes at https://github.com/hpzhang94/DiffuSum
引用
收藏
页码:13089 / 13100
页数:12
相关论文
共 44 条
  • [11] Abstractive summarization: An overview of the state of the art
    Gupta, Som
    Gupta, S. K.
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2019, 121 : 49 - 65
  • [12] Hermann K.M., 2015, Teaching machines to read and comprehend
  • [13] Ho J, 2020, Advances in neural information processing systems, P6840
  • [14] Ho Jonathan, 2022, arXiv
  • [15] Khosla P., 2020, ADV NEURAL INFORM PR, P18661, DOI [10.48550/arXiv.2004.11362, DOI 10.48550/ARXIV.2004.11362]
  • [16] Kong Z, 2020, ARXIV
  • [17] Li HR, 2020, AAAI CONF ARTIF INTE, V34, P8196
  • [18] Li Xiang Lisa, 2022, arXiv
  • [19] Lin CY, 2003, HLT-NAACL 2003: HUMAN LANGUAGE TECHNOLOGY CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, PROCEEDINGS OF THE MAIN CONFERENCE, P150
  • [20] LIU YC, 2019, ARXIV, P5238, DOI DOI 10.1109/ICCV.2019.00534