Specialized Pre-Training of Neural Networks on Synthetic Data for Improving Paraphrase Generation

被引:0
作者
O. H. Skurzhanskyi
O. O. Marchenko
A. V. Anisimov
机构
[1] Taras Shevchenko National University of Kyiv,
来源
Cybernetics and Systems Analysis | 2024年 / 60卷
关键词
artificial intelligence; machine learning; neural networks; paraphrase generation; pre-training; fine tuning;
D O I
暂无
中图分类号
学科分类号
摘要
Paraphrase generation is a fundamental problem in natural language processing. Due to the significant success of transfer learning, the “pre-training → fine-tuning” approach has become the standard. However, popular general pre-training methods typically require extensive datasets and great computational resources, and the available pre-trained models are limited by fixed architecture and size. The authors have proposed a simple and efficient approach to pre-training specifically for paraphrase generation, which noticeably improves the quality of paraphrase generation and ensures substantial enhancement of general-purpose models. They have used existing public data and new data generated by large language models. The authors have investigated how this pre-training procedure impacts neural networks of various architectures and demonstrated its efficiency across all architectures.
引用
收藏
页码:167 / 174
页数:7
相关论文
共 3 条
[1]  
Han X(2021)Pre-trained models: Past, present and future AI Open 2 225-250
[2]  
Zhang Z(undefined)undefined undefined undefined undefined-undefined
[3]  
Ding N(undefined)undefined undefined undefined undefined-undefined