Automatic Speech Transcription for Low-Resource Languages - The Case of Yoloxfochitl Mixtec (Mexico)

被引:4
|
作者
Mitral, Vikramjit [1 ]
Katholl, Andreas [1 ]
Amith, Jonathan D. [2 ]
Castillo Garcia, Rey [3 ]
机构
[1] SRI Int, Speech Technol & Res Lab, 333 Ravenswood Ave, Menlo Pk, CA 94025 USA
[2] Gettysburg Coll, Gettysburg, PA 17325 USA
[3] Secretaria Educ Publ, Chilpancingo De Los Brav, State Of Guerre, Mexico
来源
17TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2016), VOLS 1-5: UNDERSTANDING SPEECH PROCESSING IN HUMANS AND MACHINES | 2016年
基金
美国国家科学基金会;
关键词
automatic speech recognition; endangered languages; large vocabulary continuous speech recognition; articulatory features; tonal features; acoustic-phonetic features; convolutional neural networks; RECOGNITION; FEATURES;
D O I
10.21437/Interspeech.2016-546
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
The rate at which endangered languages can be documented has been highly constrained by human factors. Although digital recording of natural speech in endangered languages may proceed at a fairly robust pace, transcription of this material is not only time consuming but severely limited by the lack of native-speaker personnel proficient in the orthography of their mother tongue. Our NSF-funded project in the Documenting Endangered Languages (DEL) program proposes to tackle this problem from two sides: first via a tool that helps native speakers become proficient in the orthographic conventions of their language, and second by using automatic speech recognition (ASR) output that assists in the transcription effort for newly recorded audio data. In the present study, we focus exclusively on progress in developing speech recognition for the language of interest, Yoloxochitl Mixtec (YM), an Oto-Manguean language spoken by fewer than 5000 speakers on the Pacific coast of Guerrero, Mexico. In particular, we present results from an initial set of experiments and discuss future directions through which better and more robust acoustic models for endangered languages with limited resources can be created.
引用
收藏
页码:3076 / 3080
页数:5
相关论文
共 50 条
  • [21] The 2016 RWTH Keyword Search System for Low-Resource Languages
    Golik, Pavel
    Tueske, Zoltan
    Irie, Kazuki
    Beck, Eugen
    Schlueter, Ralf
    Ney, Hermann
    SPEECH AND COMPUTER, SPECOM 2017, 2017, 10458 : 719 - 730
  • [22] Improvement of Acoustic Models Fused with Lip Visual Information for Low-Resource Speech
    Yu, Chongchong
    Yu, Jiaqi
    Qian, Zhaopeng
    Tan, Yuchen
    SENSORS, 2023, 23 (04)
  • [23] Introducing the Urdu-Sindhi Speech Emotion Corpus: A Novel Dataset of Speech Recordings for Emotion Recognition for Two Low-Resource Languages
    Syed, Zafi Sherhan
    Memon, Sajjad Ali
    Shah, Muhammad Shehram
    Syed, Abbas Shah
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2020, 11 (04) : 805 - 810
  • [24] Convolutional Maxout Neural Networks for Low-Resource Speech Recognition
    Cai, Meng
    Shi, Yongzhe
    Kang, Jian
    Liu, Jia
    Su, Tengrong
    2014 9TH INTERNATIONAL SYMPOSIUM ON CHINESE SPOKEN LANGUAGE PROCESSING (ISCSLP), 2014, : 133 - +
  • [25] Joint Estimation of Articulatory Features and Acoustic models for Low-Resource Languages
    Abraham, Basil
    Umesh, S.
    Joy, Neethu Mariam
    18TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2017), VOLS 1-6: SITUATED INTERACTION, 2017, : 2153 - 2157
  • [26] Code-Switching ASR for Low-Resource Indic Languages: A Hindi-Marathi Case Study
    Palivela, Hemant
    Narvekar, Meera
    Asirvatham, David
    Bhushan, Shashi
    Rishiwal, Vinay
    Agarwal, Udit
    IEEE ACCESS, 2025, 13 : 9171 - 9198
  • [27] Multi-task Sequence Classification for Disjoint Tasks in Low-resource Languages
    Radom, Jarema
    Kocon, Jan
    KNOWLEDGE-BASED AND INTELLIGENT INFORMATION & ENGINEERING SYSTEMS (KSE 2021), 2021, 192 : 1132 - 1140
  • [28] Exploring the Potential of Prompting Methods in Low-Resource Speech Recognition with Whisper
    Chen, Yaqi
    Zhang, Wenlin
    Zhang, Hao
    Yang, Xukui
    Qu, Dan
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, PT III, NLPCC 2024, 2025, 15361 : 382 - 393
  • [29] Articulatory Feature based Multilingual MLPs for Low-Resource Speech Recognition
    Qian, Yanmin
    Liu, Jia
    13TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2012 (INTERSPEECH 2012), VOLS 1-3, 2012, : 2601 - 2604
  • [30] Low-resource automatic cartoon image creation from limited samples
    Cheng, Hsu-Yung
    Yu, Chih-Chang
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2020, 71