A Dialogues Summarization Algorithm Based on Multi-task Learning

被引:0
作者
Chen, Haowei [1 ]
Li, Chen [1 ]
Liang, Jiajing [1 ]
Tian, Lihua [1 ]
机构
[1] Xi An Jiao Tong Univ, Coll Software, Xian 710100, Shanxi, Peoples R China
关键词
Dialogues summarization; Multi-task; Contrast learning;
D O I
10.1007/s11063-024-11619-8
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the continuous advancement of social information, the number of texts in the form of dialogue between individuals has exponentially increased. However, it is very challenging to review the previous dialogue content before initiating a new conversation. In view of the above background, a new dialogue summarization algorithm based on multi-task learning is first proposed in the paper. Specifically, Minimum Risk Training is used as the loss function to alleviate the problem of inconsistent goals between the training phase and the testing phase. Then, in order to deal with the problem that the model cannot effectively distinguish gender pronouns, a gender pronoun discrimination auxiliary task based on contrast learning is designed to help the model learn to distinguish different gender pronouns. Finally, an auxiliary task of reducing exposure bias is introduced, which involves incorporating the summary generated during inference into another round of training to reduce the difference between the decoder inputs during the training and testing stages. Experimental results show that our model outperforms strong baselines on three public dialogue summarization datasets: SAMSUM, DialogSum, and CSDS.
引用
收藏
页数:14
相关论文
共 27 条
[1]  
Asi A., 2022, arXiv
[2]  
Chen JA, 2020, Arxiv, DOI arXiv:2010.01672
[3]  
Chen YL, 2021, Arxiv, DOI [arXiv:2105.06762, DOI 10.48550/ARXIV.2105.06762]
[4]  
Fang Y, 2022, NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, P3859
[5]  
Gao S, 2020, Arxiv, DOI arXiv:2005.04684
[6]  
Gliwa B, 2019, arXiv
[7]  
Goo CW, 2018, IEEE W SP LANG TECH, P735, DOI 10.1109/SLT.2018.8639531
[8]  
Kim S, 2022, Arxiv, DOI [arXiv:2209.00930, 10.48550/arXiv.2209.00930, DOI 10.48550/ARXIV.2209.00930]
[9]  
Lee Sebastian, 2021, P MACHINE LEARNING R, V139
[10]  
Lewis M, 2019, Arxiv, DOI [arXiv:1910.13461, 10.48550/arXiv.1910.13461]