Open-Domain Response Generation in Low-Resource Settings using Self-Supervised Pre-Training of Warm-Started Transformers

被引:4
|
作者
Naous, Tarek [1 ,3 ]
Bassyouni, Zahraa [2 ]
Mousi, Bassel [2 ]
Hajj, Hazem [2 ]
El Hajj, Wassim [2 ]
Shaban, Khaled [3 ]
机构
[1] Amer Univ Beirut, 351 Ferst Dr NW, Atlanta, GA 30332 USA
[2] Amer Univ Beirut, Bliss St,POB 11-0236, Beirut, Lebanon
[3] Qatar Univ, Univ St POB, Doha 2713, Qatar
关键词
Response generation; Arabic dialect; language models;
D O I
10.1145/3579164
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Learning response generation models constitute the main component of building open-domain dialogue systems. However, training open-domain response generation models requires large amounts of labeled data and pre-trained language generation models that are often nonexistent for low-resource languages. In this article, we propose a framework for training open-domain response generation models in low-resource settings. We consider Dialectal Arabic (DA) as a working example. The framework starts by warm-starting a transformer-based encoder-decoder with pre-trained language model parameters. Next, the resultant encoder-decoder model is adapted to DA by employing self-supervised pre-training on large-scale unlabeled data in the desired dialect. Finally, the model is fine-tuned on a very small labeled dataset for open-domain response generation. The results show significant performance improvements on three spoken Arabic dialects after adopting the framework's three stages, highlighted by higher BLEU and lower Perplexity scores compared with multiple baseline models. Specifically, our models are capable of generating fluent responses in multiple dialects with an average human-evaluated fluency score above 4. Our data is made publicly available.
引用
收藏
页数:12
相关论文
共 4 条
  • [1] Fast and Efficient Multilingual Self-Supervised Pre-training for Low-Resource Speech Recognition
    Zhang, Zhilong
    Wang, Wei
    Qian, Yanmin
    INTERSPEECH 2023, 2023, : 2248 - 2252
  • [2] Self-Supervised Audio-and-Text Pre-training with Extremely Low-Resource Parallel Data
    Kang, Yu
    Liu, Tianqiao
    Li, Hang
    Hao, Yang
    Ding, Wenbiao
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 10875 - 10883
  • [3] Self-Supervised Underwater Image Generation for Underwater Domain Pre-Training
    Wu, Zhiheng
    Wu, Zhengxing
    Chen, Xingyu
    Lu, Yue
    Yu, Junzhi
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2024, 73 : 1 - 14
  • [4] Adapter pre-training for improved speech recognition in unseen domains using low resource adapter tuning of self-supervised models
    Udupa, Sathvik
    Bandekar, Jesuraj
    Kumar, Saurabh
    Deekshitha, G.
    Sandhya, B.
    Abhayjeet, S.
    Murthy, Savitha
    Pai, Priyanka
    Raghavan, Srinivasa
    Nanavati, Raoul
    Ghosh, Prasanta Kumar
    INTERSPEECH 2024, 2024, : 2529 - 2533