Cross-Lingual Summarization Method Based on Joint Training and Self-Training in Low-Resource Scenarios

被引:0
作者
Cheng, Shaohuan [1 ]
Tang, Yujia [1 ]
Liu, Qiao [1 ]
Chen, Wenyu [1 ]
机构
[1] School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu
来源
Dianzi Keji Daxue Xuebao/Journal of the University of Electronic Science and Technology of China | 2024年 / 53卷 / 05期
关键词
cross-lingual summarization; joint training; low-resource scenarios; machine translation; self-training;
D O I
10.12178/1001-0548.2024173
中图分类号
学科分类号
摘要
As globalization continues to develop, cross-lingual summarization has become an important topic in natural language processing. In low-resource scenarios, existing methods face challenges such as limited representation transfer and insufficient data utilization. To address these issues, this paper proposes a novel method based on joint training and self-training. Specifically, two models are used to handle the translation and cross-lingual summarization tasks, respectively, which unify the language vector space of the output and avoid the issue of limited representation transfer. Additionally, joint training is performed by aligning the output features and probabilities of parallel training pairs, thereby enhancing semantic sharing between the models. Furthermore, based on joint training, a self-training technique is introduced to generate synthetic data from additional monolingual summary data, effectively mitigating the data scarcity issue of low-resource scenarios. Experimental results demonstrate that this method outperforms existing approaches in multiple low-resource scenarios, achieving significant improvements in ROUGE scores. © 2024 University of Electronic Science and Technology of China. All rights reserved.
引用
收藏
页码:762 / 770
页数:8
相关论文
共 29 条
[1]  
WANG J, MENG F, ZHENG D, Et al., A survey on cross-lingual summarization, Transactions of the Association for Computational Linguistics, 10, pp. 1304-1323, (2022)
[2]  
BAI Y, HUANG H, FAN K, Et al., Unifying cross-lingual summarization and machine translation with compression rate, Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 1087-1097, (2022)
[3]  
ZHU J, ZHOU Y, ZHANG J, Et al., Attend, translate and summarize: An efficient method for neural cross-lingual summarization, Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 1309-1321, (2020)
[4]  
LIANG Y, MENG F, ZHOU C, Et al., A variational hierarchical model for neural cross-lingual summarization, Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 2088-2099, (2022)
[5]  
BAI Y, GAO Y, HUANG H Y., Cross-lingual abstractive summarization with limited parallel resources, Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp. 6910-6924, (2021)
[6]  
NGUYEN T T, LUU A T., Improving neural cross-lingual abstractive summarization via employing optimal transport distance for knowledge distillation, Proceedings of the 36th AAAI Conference on Artificial Intelligence, pp. 11103-11111, (2022)
[7]  
ZHANG K, ZHANG Y, YU Z, Et al., A two-stage fine-tuning method for low-resource cross-lingual summarization, Mathematical Biosciences and Engineering, 21, 1, pp. 1125-1143, (2024)
[8]  
YANG X, YUN J, ZHENG B, Et al., Oversea cross-lingual summarization service in multilanguage pre-trained model through knowledge distillation, Electronics, 12, 24, (2023)
[9]  
LUO F, WANG W, LIU J, Et al., VECO: Variable and flexible cross-lingual pre-training for language understanding and generation, Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp. 3980-3994, (2021)
[10]  
THUPTEN Tsering, RINCHEN Dhondub, NYIMA Tashi, Et al., Research on Chinese-Tibetan machine translation model based on improved byte pair encoding, Journal of University of Electronic Science and Technology of China, 50, 2, pp. 249-255, (2021)