Abstractive Document Summarization via Neural Model with Joint Attention

被引:20
作者
Hou, Liwei [1 ]
Hu, Po [1 ]
Bei, Chao [2 ]
机构
[1] Cent China Normal Univ, Sch Comp Sci, Wuhan 430079, Hubei, Peoples R China
[2] Global Tone Commun Technol Co Ltd, Beijing 100043, Peoples R China
来源
NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, NLPCC 2017 | 2018年 / 10619卷
基金
中国国家自然科学基金;
关键词
Abstractive summarization; Attentional mechanism; Encoder-decoder framework; Neural network;
D O I
10.1007/978-3-319-73618-1_28
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Due to the difficulty of abstractive summarization, the great majority of past work on document summarization has been extractive, while the recent success of sequence-to-sequence framework has made abstractive summarization viable, in which a set of recurrent neural networks models based on attention encoder-decoder have achieved promising performance on short-text summarization tasks. Unfortunately, these attention encoder-decoder models often suffer from the undesirable shortcomings of generating repeated words or phrases and inability to deal with out-of-vocabulary words appropriately. To address these issues, in this work we propose to add an attention mechanism on output sequence to avoid repetitive contents and use the subword method to deal with the rare and unknown words. We applied our model to the public dataset provided by NLPCC 2017 shared task3. The evaluation results show that our system achieved the best ROUGE performance among all the participating teams and is also competitive with some state-of-the-art methods.
引用
收藏
页码:329 / 338
页数:10
相关论文
共 20 条
  • [1] [Anonymous], 2016, CONLL
  • [2] [Anonymous], 2004, ACL
  • [3] [Anonymous], 2016, NAACL
  • [4] [Anonymous], 2016, ACL
  • [5] [Anonymous], 2016, ARXIV160308148
  • [6] [Anonymous], 2014, EMNLP
  • [7] [Anonymous], 2014, 3 INT C LEARN REPR
  • [8] [Anonymous], 2016, ACL
  • [9] Erkan Gunes., 2004, EMNLP
  • [10] Gillick D, 2009, ACL