A Joint Learning Framework With BERT for Spoken Language Understanding

被引:23
作者
Zhang, Zhichang [1 ]
Zhang, Zhenwen [1 ]
Chen, Haoyuan [1 ]
Zhang, Zhiman [1 ]
机构
[1] Northwest Normal Univ, Coll Comp Sci & Engn, Lanzhou 730070, Peoples R China
来源
IEEE ACCESS | 2019年 / 7卷
基金
中国国家自然科学基金;
关键词
Spoken language understanding; intent classification and slot filling; joint learning; intent-augmented mechanism; pre-trained language model;
D O I
10.1109/ACCESS.2019.2954766
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Intent classification and slot filling are two essential tasks for spoken language understanding. Recently, joint learning has been shown to be effective for the two tasks. However, most joint learning methods only consider joint learning using shared parameters on the surface level rather than the semantic level, and these methods suffer from small-scale human-labeled training data, resulting in poor generalization capabilities, especially for rare words. In this paper, we propose a novel encoder-decoder framework based multi-task learning model, which conducts joint training for intent classification and slot filling tasks. For the encoder of our model, we encode the input sequence as context representations using bidirectional encoder representation from transformers (BERT). For the decoder, we implement two-stage decoder process in our model. In the first stage, we use an intent classification decoder to detect the user's intent. In the second stage, we leverage the intent contextual information into the slot filling decoder to predict the semantic concept tags for each word. We conduct experiments on three popular benchmark datasets: ATIS, Snips and Facebook multilingual task-oriented datasets. The experimental results show that our proposed model outperforms the state-of-the-art approaches and achieves new state-of-the-art results on both three datasets.
引用
收藏
页码:168849 / 168858
页数:10
相关论文
共 45 条
[1]  
[Anonymous], P 8 ANN C INT SPEECH
[2]  
[Anonymous], 2016, P C ASS MACH TRANSL
[3]  
[Anonymous], 2014, 3 INT C LEARN REPR
[4]  
[Anonymous], P NAACL
[5]  
Bojanowski Piotr, 2017, Trans. Assoc. Comput. Linguist., V5, P135, DOI DOI 10.1162/TACL_A_00051
[6]  
Chen Q., 2019, ARXIV190210909
[7]  
Coucke A., 2018, ARXIV180510190 818
[8]  
Deoras A, 2013, INTERSPEECH, P2712
[9]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[10]  
E HH, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), P5467