AI-based abstractive text summarization towards AIoT and edge computing

被引:1
作者
Ma, Jun [1 ]
Li, Tong [2 ]
Zhang, Yanling [2 ]
机构
[1] China Mobile Grp Shaanxi Co Ltd, Xian, Shaanxi, Peoples R China
[2] Xidian Univ, State Key Lab Integrated Serv, 2 Taibai South Rd, Xian, Shaanxi, Peoples R China
关键词
abstractive summarization; AIoT and edge computing; depth-wise separable convolution; gated linear unit; self-attention; transformer;
D O I
10.1002/itl2.354
中图分类号
TN [电子技术、通信技术];
学科分类号
0809 ;
摘要
The task of text summarization has been widely concerned in the fields of news, document indexing and literature retrieval. Recently, due to the rise of mobile Internet devices, natural language processing for AIoT and edge computing has become a hot spot. This paper focuses on the research of text summarization for AIoT and edge computing. For a long time, abstractive summarization is limited to academic research due to the lack of controllability of the generated content. Recently, the appearance of Transformer has changed the current situation of abstractive text summarization. Transformer follows encoder-decoder architecture, including attention mechanism and feed-forward network. The encoder encodes the semantic information of source text, and decoder adaptively selects the effective context information through the attention mechanism to generate a coherent summary. To extract more semantic information and control the generated text better, this paper proposes multi-scale semantic information Transformer (MSIT). Specifically, we introduce depth-wise separable convolution to the encoder to extract more local semantic information, so that the attention mechanism can make better use of contextual semantic information. Additionally, we combine the encoding vector of encoder and target summary as the input to the attention layer of the decoder, and introduce time series mechanism so that the decoder can consider context information when generating text. Experiments on CNN-Daily Mail Dataset show that this model is superior to other methods.
引用
收藏
页数:6
相关论文
共 14 条
  • [1] Allahyari M, 2017, INT J ADV COMPUT SC, V8, P397, DOI 10.14569/IJACSA.2017.081052
  • [2] Cho K., 2014, P SSST 8 8 WORKSHOP, DOI 10.3115/v1/w14-4012
  • [3] Xception: Deep Learning with Depthwise Separable Convolutions
    Chollet, Francois
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 1800 - 1807
  • [4] Dauphin YN, 2017, PR MACH LEARN RES, V70
  • [5] Recent automatic text summarization techniques: a survey
    Gambhir, Mahak
    Gupta, Vishal
    [J]. ARTIFICIAL INTELLIGENCE REVIEW, 2017, 47 (01) : 1 - 66
  • [6] Gehrmann S, 2018, 2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), P4098
  • [7] Joulin A., 2016, Bag of tricks for efficient text classification
  • [8] Shaw P., 2018, P 2018 C N AM CHAPTE, P464, DOI DOI 10.18653/V1/N18-2074
  • [9] Sutskever I, 2014, ADV NEUR IN, V27
  • [10] Rethinking the Inception Architecture for Computer Vision
    Szegedy, Christian
    Vanhoucke, Vincent
    Ioffe, Sergey
    Shlens, Jon
    Wojna, Zbigniew
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 2818 - 2826