Automatic Speech Transcription for Low-Resource Languages - The Case of Yoloxfochitl Mixtec (Mexico)

被引:4
|
作者
Mitral, Vikramjit [1 ]
Katholl, Andreas [1 ]
Amith, Jonathan D. [2 ]
Castillo Garcia, Rey [3 ]
机构
[1] SRI Int, Speech Technol & Res Lab, 333 Ravenswood Ave, Menlo Pk, CA 94025 USA
[2] Gettysburg Coll, Gettysburg, PA 17325 USA
[3] Secretaria Educ Publ, Chilpancingo De Los Brav, State Of Guerre, Mexico
来源
17TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2016), VOLS 1-5: UNDERSTANDING SPEECH PROCESSING IN HUMANS AND MACHINES | 2016年
基金
美国国家科学基金会;
关键词
automatic speech recognition; endangered languages; large vocabulary continuous speech recognition; articulatory features; tonal features; acoustic-phonetic features; convolutional neural networks; RECOGNITION; FEATURES;
D O I
10.21437/Interspeech.2016-546
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
The rate at which endangered languages can be documented has been highly constrained by human factors. Although digital recording of natural speech in endangered languages may proceed at a fairly robust pace, transcription of this material is not only time consuming but severely limited by the lack of native-speaker personnel proficient in the orthography of their mother tongue. Our NSF-funded project in the Documenting Endangered Languages (DEL) program proposes to tackle this problem from two sides: first via a tool that helps native speakers become proficient in the orthographic conventions of their language, and second by using automatic speech recognition (ASR) output that assists in the transcription effort for newly recorded audio data. In the present study, we focus exclusively on progress in developing speech recognition for the language of interest, Yoloxochitl Mixtec (YM), an Oto-Manguean language spoken by fewer than 5000 speakers on the Pacific coast of Guerrero, Mexico. In particular, we present results from an initial set of experiments and discuss future directions through which better and more robust acoustic models for endangered languages with limited resources can be created.
引用
收藏
页码:3076 / 3080
页数:5
相关论文
共 50 条
  • [31] Domain Adaptation Speech-to-Text for Low-Resource European Portuguese Using Deep Learning
    Medeiros, Eduardo
    Corado, Leonel
    Rato, Luis
    Quaresma, Paulo
    Salgueiro, Pedro
    FUTURE INTERNET, 2023, 15 (05)
  • [32] ScoutWav: Two-Step Fine-Tuning on Self-Supervised Automatic Speech Recognition for Low-Resource Environments
    Fatehi, Kavan
    Torres, Mercedes Torres
    Kucukyilmaz, Ayse
    INTERSPEECH 2022, 2022, : 3523 - 3527
  • [33] Feature learning for efficient ASR-free keyword spotting in low-resource languages
    van der Westhuizen, Ewald
    Kamper, Herman
    Menon, Raghav
    Quinn, John
    Niesler, Thomas
    COMPUTER SPEECH AND LANGUAGE, 2022, 71
  • [34] Acoustic Modeling Based on Deep Learning for Low-Resource Speech Recognition: An Overview
    Yu, Chongchong
    Kang, Meng
    Chen, Yunbing
    Wu, Jiajia
    Zhao, Xia
    IEEE ACCESS, 2020, 8 : 163829 - 163843
  • [35] Development of a low-resource wearable continuous gesture-to-speech conversion system
    Parthasarathy, Vijayalakshmi
    Thangavelu, Nagarajan
    Ramesh, Jayapriya
    Suresh, Brathindara
    Kandasamy, Krithika
    Nikhilesh, N.
    Nagarajan, Narenraju
    Sathyasingh, Johanan Joysingh
    Vijayakumar, Aiswarya
    Kannan, Mrinalini
    DISABILITY AND REHABILITATION-ASSISTIVE TECHNOLOGY, 2023, 18 (08) : 1441 - 1452
  • [36] LEARNING FROM THE BEST: A TEACHER-STUDENT MULTILINGUAL FRAMEWORK FOR LOW-RESOURCE LANGUAGES
    Bagchi, Deblin
    Hartmann, William
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 6051 - 6055
  • [37] Lightweight Automatic Modulation Classification Based on Efficient Convolution and Graph Sparse Attention in Low-Resource Scenarios
    Cai, Zhuoran
    Wang, Chuan
    Ma, Wenxuan
    Li, Xiangzhen
    Zhou, Ruoyu
    IEEE INTERNET OF THINGS JOURNAL, 2025, 12 (04): : 3629 - 3638
  • [38] A Semi-Supervised Complementary Joint Training Approach for Low-Resource Speech Recognition
    Du, Ye-Qian
    Zhang, Jie
    Fang, Xin
    Wu, Ming-Hui
    Yang, Zhou-Wang
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2023, 31 : 3908 - 3921
  • [39] IMPROVING THE PERFORMANCE OF TRANSFORMER BASED LOW RESOURCE SPEECH RECOGNITION FOR INDIAN LANGUAGES
    Shetty, Vishwas M.
    Mary, Metilda Sagaya N. J.
    Umesh, S.
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 8279 - 8283
  • [40] TAML-Adapter: Enhancing Adapter Tuning Through Task-Agnostic Meta-Learning for Low-Resource Automatic Speech Recognition
    Liu, Yunpeng
    Yang, Xukui
    Zhang, Jiayi
    Xi, Yangli
    Qu, Dan
    IEEE SIGNAL PROCESSING LETTERS, 2025, 32 : 636 - 640