A Semantic Supervision Method for Abstractive Summarization

被引:4
作者
Hu, Sunqiang [1 ]
Li, Xiaoyu [1 ]
Deng, Yu [1 ]
Peng, Yu [1 ]
Lin, Bin [2 ]
Yang, Shan [3 ]
机构
[1] Univ Elect Sci & Technol China, Sch Informat & Software Engn, Chengdu 610054, Peoples R China
[2] Sichuan Normal Univ, Sch Engn, Chengdu 610066, Peoples R China
[3] Jackson State Univ, Dept Chem Phys & Atmospher Sci, Jackson, MS 39217 USA
来源
CMC-COMPUTERS MATERIALS & CONTINUA | 2021年 / 69卷 / 01期
基金
中国国家自然科学基金; 国家重点研发计划;
关键词
Text summarization; semantic supervision; capsule network;
D O I
10.32604/cmc.2021.017441
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In recent years, many text summarization models based on pretraining methods have achieved very good results. However, in these text summarization models, semantic deviations are easy to occur between the original input representation and the representation that passed multi-layer encoder, which may result in inconsistencies between the generated summary and the source text content. The Bidirectional Encoder Representations from Transformers (BERT) improves the performance of many tasks in Natural Language Processing (NLP). Although BERT has a strong capability to encode context, it lacks the fine-grained semantic representation. To solve these two problems, we proposed a semantic supervision method based on Capsule Network. Firstly, we extracted the fine-grained semantic representation of the input and encoded result in BERT by Capsule Network. Secondly, we used the fine-grained semantic representation of the input to supervise the fine-grained semantic representation of the encoded result. Then we evaluated our model on a popular Chinese social media dataset (LCSTS), and the result showed that our model achieved higher ROUGE scores (including R-1, R-2), and our model outperformed baseline systems. Finally, we conducted a comparative study on the stability of the model, and the experimental results showed that our model was more stable.
引用
收藏
页码:145 / 158
页数:14
相关论文
共 27 条
[11]  
Ioffe S, 2015, PR MACH LEARN RES, V37, P448
[12]  
King DB, 2015, ACS SYM SER, V1214, P1, DOI 10.1021/bk-2015-1214.ch001
[13]  
Koehn P, 2004, LECT NOTES COMPUT SC, V3265, P115
[14]  
Li Piji, 2017, P 2017 C EMP METH NA, P2091, DOI DOI 10.18653/V1/D17-1222
[15]  
Lin C.Y., 2004, WORKSHOP TEXT SUMMAR, P74, DOI [10.3115/1076090.1076095, DOI 10.1253/JCJ.34.1213]
[16]  
Liu Y, 2019, 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019), P3730
[17]  
Ma S., 2018, PROC NAACL HLT, P196, DOI [10.18653/v1/N18-1018, DOI 10.18653/V1]
[18]  
Ma SM, 2018, PROCEEDINGS OF THE 56TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 2, P725
[19]   Improving Semantic Relevance for Sequence-to-Sequence Learning of Chinese Social Media Text Summarization [J].
Ma, Shuming ;
Sun, Xu ;
Xu, Jingjing ;
Wang, Houfeng ;
Li, Wenjie ;
Su, Qi .
PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 2, 2017, :635-640
[20]  
Peters M.E., 2018, P NAACL HLT 2018 NEW, P2227