Incorporating word attention with convolutional neural networks for abstractive summarization

被引:0
|
作者
Chengzhe Yuan
Zhifeng Bao
Mark Sanderson
Yong Tang
机构
[1] South China Normal University,School of Computer Science
[2] RMIT University,School of Science, Computer Science and Information Technology
来源
World Wide Web | 2020年 / 23卷
关键词
Abstractive summarization; Word attention; Convolutional neural networks; Sequence-to-sequence model;
D O I
暂无
中图分类号
学科分类号
摘要
Neural sequence-to-sequence (seq2seq) models have been widely used in abstractive summarization tasks. One of the challenges of this task is redundant contents in the input document often confuses the models and leads to poor performance. An efficient way to solve this problem is to select salient information from the input document. In this paper, we propose an approach that incorporates word attention with multilayer convolutional neural networks (CNNs) to extend a standard seq2seq model for abstractive summarization. First, by concentrating on a subset of source words during encoding an input sentence, word attention is able to extract informative keywords in the input, which gives us the ability to interpret generated summaries. Second, these keywords are further distilled by multilayer CNNs to capture the coarse-grained contextual features of the input sentence. Thus, the combined word attention and multilayer CNNs modules provide a better-learned representation of the input document, which helps the model generate interpretable, coherent and informative summaries in an abstractive summarization task. We evaluate the effectiveness of our model on the English Gigaword, DUC2004 and Chinese summarization dataset LCSTS. Experimental results show the effectiveness of our approach.
引用
收藏
页码:267 / 287
页数:20
相关论文
共 50 条
  • [31] Brain-model neural similarity reveals abstractive summarization performance
    Zhang, Zhejun
    Guo, Shaoting
    Zhou, Wenqing
    Luo, Yingying
    Zhu, Yingqi
    Zhang, Lin
    Li, Lei
    SCIENTIFIC REPORTS, 2025, 15 (01):
  • [32] Towards Sequence-to-Sequence Neural Model for Croatian Abstractive Summarization
    Davidovic, Vlatka
    Ipsic, Sanda Martincic
    CENTRAL EUROPEAN CONFERENCE ON INFORMATION AND INTELLIGENT SYSTEMS, CECIIS, 2023, : 309 - 315
  • [33] Enhancing Abstractive Summarization with Pointer Generator Networks and Coverage Mechanisms in NLP
    Yarlagadda, Madhulika
    Nadendla, Hanumantha Rao
    IRANIAN JOURNAL OF SCIENCE AND TECHNOLOGY-TRANSACTIONS OF ELECTRICAL ENGINEERING, 2024, 48 (03) : 1301 - 1311
  • [34] Dense & Attention Convolutional Neural Networks for Toe Walking Recognition
    Chen, Junde
    Soangra, Rahul
    Grant-Beuttler, Marybeth
    Nanehkaran, Y. A.
    Wen, Yuxin
    IEEE TRANSACTIONS ON NEURAL SYSTEMS AND REHABILITATION ENGINEERING, 2023, 31 : 2235 - 2245
  • [35] Multiscale Hybrid Convolutional Deep Neural Networks with Channel Attention
    Yang, Hua
    Yang, Ming
    He, Bitao
    Qin, Tao
    Yang, Jing
    ENTROPY, 2022, 24 (09)
  • [36] A Convolution-Self Attention Abstractive Summarization Method Fusing Sequential Grammar Knowledge
    Luo S.
    Wang R.
    Wu Q.
    Pan L.
    Wu Z.
    Beijing Ligong Daxue Xuebao/Transaction of Beijing Institute of Technology, 2021, 41 (01): : 93 - 101
  • [37] BCA: Bilinear Convolutional Neural Networks and Attention Networks for legal question answering
    Zhang, Haiguang
    Zhang, Tongyue
    Cao, Faxin
    Wang, Zhizheng
    Zhang, Yuanyu
    Sun, Yuanyuan
    Vicente, Mark Anthony
    AI OPEN, 2022, 3 : 172 - 181
  • [38] A Simple and Light-Weight Attention Module for Convolutional Neural Networks
    Park, Jongchan
    Woo, Sanghyun
    Lee, Joon-Young
    Kweon, In So
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2020, 128 (04) : 783 - 798
  • [39] Attention-Aware Multi-Task Convolutional Neural Networks
    Lyu, Kejie
    Li, Yingming
    Zhang, Zhongfei
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 : 1867 - 1878
  • [40] A Simple and Light-Weight Attention Module for Convolutional Neural Networks
    Jongchan Park
    Sanghyun Woo
    Joon-Young Lee
    In So Kweon
    International Journal of Computer Vision, 2020, 128 : 783 - 798