Incorporating word attention with convolutional neural networks for abstractive summarization

被引:0
|
作者
Chengzhe Yuan
Zhifeng Bao
Mark Sanderson
Yong Tang
机构
[1] South China Normal University,School of Computer Science
[2] RMIT University,School of Science, Computer Science and Information Technology
来源
World Wide Web | 2020年 / 23卷
关键词
Abstractive summarization; Word attention; Convolutional neural networks; Sequence-to-sequence model;
D O I
暂无
中图分类号
学科分类号
摘要
Neural sequence-to-sequence (seq2seq) models have been widely used in abstractive summarization tasks. One of the challenges of this task is redundant contents in the input document often confuses the models and leads to poor performance. An efficient way to solve this problem is to select salient information from the input document. In this paper, we propose an approach that incorporates word attention with multilayer convolutional neural networks (CNNs) to extend a standard seq2seq model for abstractive summarization. First, by concentrating on a subset of source words during encoding an input sentence, word attention is able to extract informative keywords in the input, which gives us the ability to interpret generated summaries. Second, these keywords are further distilled by multilayer CNNs to capture the coarse-grained contextual features of the input sentence. Thus, the combined word attention and multilayer CNNs modules provide a better-learned representation of the input document, which helps the model generate interpretable, coherent and informative summaries in an abstractive summarization task. We evaluate the effectiveness of our model on the English Gigaword, DUC2004 and Chinese summarization dataset LCSTS. Experimental results show the effectiveness of our approach.
引用
收藏
页码:267 / 287
页数:20
相关论文
共 50 条
  • [21] Neural abstractive summarization fusing by global generative topics
    Gao, Yang
    Wang, Yang
    Liu, Luyang
    Guo, Yidi
    Huang, Heyan
    NEURAL COMPUTING & APPLICATIONS, 2020, 32 (09) : 5049 - 5058
  • [22] Abstractive Text Summarization Using Pointer-Generator Networks With Pre-trained Word Embedding
    Dang Trung Anh
    Nguyen Thi Thu Trang
    SOICT 2019: PROCEEDINGS OF THE TENTH INTERNATIONAL SYMPOSIUM ON INFORMATION AND COMMUNICATION TECHNOLOGY, 2019, : 473 - 478
  • [23] Legal public opinion news abstractive summarization by incorporating topic information
    Yuxin Huang
    Zhengtao Yu
    Junjun Guo
    Zhiqiang Yu
    Yantuan Xian
    International Journal of Machine Learning and Cybernetics, 2020, 11 : 2039 - 2050
  • [24] Legal public opinion news abstractive summarization by incorporating topic information
    Huang, Yuxin
    Yu, Zhengtao
    Guo, Junjun
    Yu, Zhiqiang
    Xian, Yantuan
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2020, 11 (09) : 2039 - 2050
  • [25] HITS-based attentional neural model for abstractive summarization
    Cai, Xiaoyan
    Shi, Kaile
    Jiang, Yuehan
    Yang, Libin
    Liu, Sen
    KNOWLEDGE-BASED SYSTEMS, 2021, 222
  • [26] Probabilistic Attention Map: A Probabilistic Attention Mechanism for Convolutional Neural Networks
    Liu, Yifeng
    Tian, Jing
    SENSORS, 2024, 24 (24)
  • [27] Spatial Decomposition and Aggregation for Attention in Convolutional Neural Networks
    Zhu, Meng
    Min, Weidong
    Xiang, Hongyue
    Zha, Cheng
    Huang, Zheng
    Li, Longfei
    Fu, Qiyan
    INTERNATIONAL JOURNAL OF PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE, 2024, 38 (01)
  • [28] An Abstractive Summarization Model Based on Joint-Attention Mechanism and a Priori Knowledge
    Li, Yuanyuan
    Huang, Yuan
    Huang, Weijian
    Yu, Junhao
    Huang, Zheng
    APPLIED SCIENCES-BASEL, 2023, 13 (07):
  • [29] Handwritten English Word Recognition based on Convolutional Neural Networks
    Yuan, Aiquan
    Bai, Gang
    Yang, Po
    Guo, Yanni
    Zhao, Xinting
    13TH INTERNATIONAL CONFERENCE ON FRONTIERS IN HANDWRITING RECOGNITION (ICFHR 2012), 2012, : 207 - 212
  • [30] KAAS: A Keyword-Aware Attention Abstractive Summarization Model for Scientific Articles
    Li, Shuaimin
    Xu, Jungang
    DATABASE SYSTEMS FOR ADVANCED APPLICATIONS, DASFAA 2022, PT III, 2022, : 263 - 271