COMPARISON OF SELF-SUPERVISED SPEECH PRE-TRAINING METHODS ON FLEMISH DUTCH

被引:1
作者
Poncelet, Jakob [1 ]
Hamme, Hugo Van [1 ]
机构
[1] Katholieke Univ Leuven, Dept Elect Engn ESAT PSI, Kasteelpk Arenberg 10,Bus 2441, B-3001 Leuven, Belgium
来源
2021 IEEE AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING WORKSHOP (ASRU) | 2021年
关键词
speech recognition; self-supervised learning; pre-training; cross-lingual;
D O I
10.1109/ASRU51503.2021.9688061
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent research in speech processing exhibits a growing interest in unsupervised and self-supervised representation learning from unlabelled data to alleviate the need for large amounts of annotated data. We investigate several popular pre-training methods and apply them to Flemish Dutch. We compare off-the-shelf English pre-trained models to models trained on an increasing amount of Flemish data. We find that the most important factors for positive transfer to downstream speech recognition tasks include a substantial amount of data and a matching pre-training domain. Ideally, we also finetune on an annotated subset in the target language. All pre-trained models improve linear phone separability in Flemish, but not all methods improve Automatic Speech Recognition. We experience superior performance with wav2vec 2.0 and we obtain a 30% WER improvement by finetuning the multilingually pre-trained XLSR-53 model on Flemish Dutch, after integration into an HMM-DNN acoustic model.
引用
收藏
页码:169 / 176
页数:8
相关论文
共 50 条
  • [41] Abdominal Organs and Pan-Cancer Segmentation Based on Self-supervised Pre-training and Self-training
    Li, He
    Han, Meng
    Wang, Guotai
    FAST, LOW-RESOURCE, AND ACCURATE ORGAN AND PAN-CANCER SEGMENTATION IN ABDOMEN CT, FLARE 2023, 2024, 14544 : 130 - 142
  • [42] SDCluster: A clustering based self-supervised pre-training method for semantic segmentation of remote sensing images
    Xu, Hanwen
    Zhang, Chenxiao
    Yue, Peng
    Wang, Kaixuan
    ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2025, 223 : 1 - 14
  • [43] Single-atom catalysts property prediction via Supervised and Self-Supervised pre-training models
    Wang, Lanjing
    Chen, Honghao
    Yang, Longqi
    Li, Jiali
    Li, Yong
    Wang, Xiaonan
    CHEMICAL ENGINEERING JOURNAL, 2024, 487
  • [44] Self-Supervised Pre-Training with Bridge Neural Network for SAR-Optical Matching
    Qian, Lixin
    Liu, Xiaochun
    Huang, Meiyu
    Xiang, Xueshuang
    REMOTE SENSING, 2022, 14 (12)
  • [45] Self-Supervised Pre-Training for Deep Image Prior-Based Robust PET Image Denoising
    Onishi, Yuya
    Hashimoto, Fumio
    Ote, Kibo
    Matsubara, Keisuke
    Ibaraki, Masanobu
    IEEE TRANSACTIONS ON RADIATION AND PLASMA MEDICAL SCIENCES, 2024, 8 (04) : 348 - 356
  • [46] Robust wav2vec 2.0: Analyzing Domain Shift in Self-Supervised Pre-Training
    Hsu, Wei-Ning
    Sriram, Anuroop
    Baevski, Alexei
    Likhomanenko, Tatiana
    Xu, Qiantong
    Pratap, Vineel
    Kahn, Jacob
    Lee, Ann
    Collobert, Ronan
    Synnaeve, Gabriel
    Auli, Michael
    INTERSPEECH 2021, 2021, : 721 - 725
  • [47] PerFedRec plus plus : Enhancing Personalized Federated Recommendation with Self-Supervised Pre-Training
    Luo, Sichun
    Xiao, Yuanzhang
    Zhang, Xinyi
    Liu, Yang
    Ding, Wenbo
    Song, Linqi
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2024, 15 (05)
  • [48] Self-Supervised Pre-Training for 3-D Roof Reconstruction on LiDAR Data
    Yang, Hongxin
    Huang, Shangfeng
    Wang, Ruisheng
    Wang, Xin
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2024, 21 : 1 - 5
  • [49] Self-supervised multimodal reconstruction pre-training for retinal computer-aided diagnosis
    Hervella, Alvaro S.
    Rouco, Jose
    Novo, Jorge
    Ortega, Marcos
    EXPERT SYSTEMS WITH APPLICATIONS, 2021, 185
  • [50] Multi-modal cross-domain self-supervised pre-training for fMRI and EEG fusion
    Wei, Xinxu
    Zhao, Kanhao
    Jiao, Yong
    Carlisle, Nancy B.
    Xie, Hua
    Fonzo, Gregory A.
    Zhang, Yu
    NEURAL NETWORKS, 2025, 184