Improving Transformer with Sequential Context Representations for Abstractive Text Summarization

被引:21
|
作者
Cai, Tian [1 ,2 ]
Shen, Mengjun [1 ,2 ]
Peng, Huailiang [1 ,2 ]
Jiang, Lei [1 ]
Dai, Qiong [1 ]
机构
[1] Chinese Acad Sci, Inst Informat Engn, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Sch Cyber Secur, Beijing, Peoples R China
来源
NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING (NLPCC 2019), PT I | 2019年 / 11838卷
基金
美国国家科学基金会;
关键词
Transformer; Abstractive summarization;
D O I
10.1007/978-3-030-32233-5_40
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent dominant approaches for abstractive text summarization are mainly RNN-based encoder-decoder framework, these methods usually suffer from the poor semantic representations for long sequences. In this paper, we propose a new abstractive summarization model, called RC-Transformer (RCT). The model is not only capable of learning longterm dependencies, but also addresses the inherent shortcoming of Transformer on insensitivity to word order information. We extend the Transformer with an additional RNN-based encoder to capture the sequential context representations. In order to extract salient information effectively, we further construct a convolution module to filter the sequential context with local importance. The experimental results on Gigaword and DUC-2004 datasets show that our proposed model achieves the state-of-the-art performance, even without introducing external information. In addition, our model also owns an advantage in speed over the RNN-based models.
引用
收藏
页码:512 / 524
页数:13
相关论文
共 50 条
  • [31] English-Arabic Text Translation and Abstractive Summarization Using Transformers
    Holiel, Heidi Ahmed
    Mohamed, Nancy
    Ahmed, Arwa
    Medhat, Walaa
    2023 20TH ACS/IEEE INTERNATIONAL CONFERENCE ON COMPUTER SYSTEMS AND APPLICATIONS, AICCSA, 2023,
  • [32] TASP : Topic-based abstractive summarization of Facebook text posts
    Benedetto, Irene
    La Quatra, Moreno
    Cagliero, Luca
    Vassio, Luca
    Trevisan, Martino
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 255
  • [33] CATAMARAN: A Cross-lingual Long Text Abstractive Summarization Dataset
    Chen, Zheng
    Lin, Hongyu
    LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2022, : 6932 - 6937
  • [34] Anaphora resolved abstractive text summarization (AR-ATS) system
    Moratanch, N.
    Chitrakala, S.
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 82 (03) : 4569 - 4597
  • [35] A Survey of the State-of-the-Art Models in Neural Abstractive Text Summarization
    Syed, Ayesha Ayub
    Gaol, Ford Lumban
    Matsuo, Tokuro
    IEEE ACCESS, 2021, 9 : 13248 - 13265
  • [36] Anaphora resolved abstractive text summarization (AR-ATS) system
    N. Moratanch
    S. Chitrakala
    Multimedia Tools and Applications, 2023, 82 : 4569 - 4597
  • [37] Neural Attention Model for Abstractive Text Summarization Using Linguistic Feature Space
    Dilawari, Aniqa
    Khan, Muhammad Usman Ghani
    Saleem, Summra
    Zahoor-Ur-Rehman
    Shaikh, Fatema Sabeen
    IEEE ACCESS, 2023, 11 : 23557 - 23564
  • [38] Abstractive text summarization of low- resourced languages using deep learning
    Shafiq, Nida
    Hamid, Isma
    Asif, Muhammad
    Nawaz, Qamar
    Aljuaid, Hanan
    Ali, Hamid
    PEERJ COMPUTER SCIENCE, 2023, 9
  • [39] A Context-Aware BERT Retrieval Framework Utilizing Abstractive Summarization
    Pan, Min
    Li, Teng
    Yang, Chenghao
    Zhou, Shuting
    Feng, Shaoxiong
    Fang, Youbin
    Li, Xingyu
    2022 IEEE/WIC/ACM INTERNATIONAL JOINT CONFERENCE ON WEB INTELLIGENCE AND INTELLIGENT AGENT TECHNOLOGY, WI-IAT, 2022, : 873 - 878
  • [40] Unified extractive-abstractive summarization: a hybrid approach utilizing BERT and transformer models for enhanced document summarization
    Divya, S.
    Sripriya, N.
    Andrew, J.
    Mazzara, Manuel
    PEERJ COMPUTER SCIENCE, 2024, 10 : 1 - 26