Multi-Level Cross-Lingual Transfer Learning With Language Shared and Specific Knowledge for Spoken Language Understanding

被引:7
作者
He, Keqing [1 ]
Xu, Weiran [1 ]
Yan, Yuanmeng [1 ]
机构
[1] Beijing Univ Posts & Telecommun, Dept Informat & Commun Engn, Beijing 100876, Peoples R China
来源
IEEE ACCESS | 2020年 / 8卷
关键词
Spoken language understanding; cross-lingual learning; linguistic knowledge transfer; adversarial learning; multi-level knowledge representation;
D O I
10.1109/ACCESS.2020.2972925
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Recently conversational agents effectively improve their understanding capabilities by neural networks. Such deep neural models, however, do not apply to most human languages due to the lack of annotated training data for various NLP tasks. In this paper, we propose a multi-level cross-lingual transfer model with language shared and specific knowledge to improve the spoken language understanding of low-resource languages. Our method explicitly separates the model into the language-shared part and language-specific part to transfer cross-lingual knowledge and improve the monolingual slot tagging, especially for low-resource languages. To refine the shared knowledge, we add a language discriminator and employ adversarial training to reinforce information separation. Besides, we adopt novel multi-level knowledge transfer in an incremental and progressive way to acquire multi-granularity shared knowledge rather than a single layer. To mitigate the discrepancies between the feature distributions of language specific and shared knowledge, we propose the neural adapters to fuse knowledge automatically. Experiments show that our proposed model consistently outperforms monolingual baseline with a statistically significant margin up to 2.09;, even higher improvement of 12.21; in the zero-shot setting.
引用
收藏
页码:29407 / 29416
页数:10
相关论文
共 40 条
[1]  
[Anonymous], 2018, ARXIV180506297
[2]  
[Anonymous], 2017, ARXIV170506273
[3]  
[Anonymous], 2016, 15 ANN C N AM CHAPT
[4]  
[Anonymous], 2015, P 2015 INT C N AM CH
[5]  
[Anonymous], 2017, ICLR 2017
[6]  
[Anonymous], 2016, NEURAL INFORM PROCES
[7]  
[Anonymous], [No title captured]
[8]  
Chen L., 2019, ARXIV190205309
[9]  
Chiu J., 2016, Trans. Assoc. Comput. Linguist., V4, P357
[10]  
Conneau A., 2017, ARXIV171004087