Toward Zero-Shot and Zero-Resource Multilingual Question Answering

被引:3
作者
Kuo, Chia-Chih [1 ]
Chen, Kuan-Yu [1 ]
机构
[1] Natl Taiwan Univ Sci & Technol, Comp Sci & Informat Engn Dept, Taipei 106, Taiwan
关键词
Task analysis; Question answering (information retrieval); Training data; Data models; Transfer learning; Online services; Internet; Natural language processing; Multilingual question answering; zero-shot; zero-resource; mBERT;
D O I
10.1109/ACCESS.2022.3207569
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In recent years, multilingual question answering has been an emergent research topic and has attracted much attention. Although systems for English and other rich-resource languages that rely on various advanced deep learning-based techniques have been highly developed, most of them in low-resource languages are impractical due to data insufficiency. Accordingly, many studies have attempted to improve the performance of low-resource languages in a zero-shot or few-shot manner based on multilingual bidirectional encoder representations from transformers (mBERT) by transferring knowledge learned from rich-resource languages to low-resource languages. Most methods require either a large amount of unlabeled data or a small set of labeled data for low-resource languages. In Wikipedia, 169 languages have less than 10,000 articles, and 48 languages have less than 1,000 articles. This reason motivates us to conduct a zero-shot multilingual question answering task under a zero-resource scenario. Thus, this study proposes a framework to fine-tune the original mBERT using data from rich-resource languages, and the resulting model can be used for low-resource languages in a zero-shot and zero-resource manner. Compared to several baseline systems, which require millions of unlabeled data for low-resource languages, the performance of our proposed framework is not only highly comparative but is also better for languages used in training.
引用
收藏
页码:99754 / 99761
页数:8
相关论文
共 38 条
  • [1] Massively Multilingual Sentence Embeddings for Zero-Shot Cross-Lingual Transfer and Beyond
    Artetxe, Mikel
    Schwenk, Holger
    [J]. TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2019, 7 : 597 - 610
  • [2] A neural probabilistic language model
    Bengio, Y
    Ducharme, R
    Vincent, P
    Jauvin, C
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2003, 3 (06) : 1137 - 1155
  • [3] Chi Z., 2020, ARXIV
  • [4] Chi ZW, 2022, Arxiv, DOI arXiv:2106.16138
  • [5] TYDI QA: A Benchmark for Information-Seeking Question Answering in Typologically Diverse Languages
    Clark, Jonathan H.
    Choi, Eunsol
    Collins, Michael
    Garrette, Dan
    Kwiatkowski, Tom
    Nikolaev, Vitaly
    Palomaki, Jennimaria
    [J]. TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2020, 8 : 454 - 470
  • [6] ELECTRA: PRE-TRAINING TEXT ENCODERS AS DISCRIMINATORS RATHER THAN GENERATORS
    Clark, Kevin
    Luong, Minh-Thang
    Le, Quoc V.
    Manning, Christopher D.
    [J]. INFORMATION SYSTEMS RESEARCH, 2020,
  • [7] Conneau A., 2020, P 58 ANN M ASS COMP, P8440
  • [8] Conneau A, 2018, Arxiv, DOI arXiv:1710.04087
  • [9] Conneau Alexis., 2018, 6 INT C LEARNING REP
  • [10] Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171