A Multi-domain Text Classification Method Based on Recurrent Convolution Multi-task Learning

被引:3
|
作者
Xie Jinbao [1 ]
Li Jiahui [2 ]
Kang Shouqiang [2 ]
Wang Qingyan [2 ]
Wang Yujing [2 ]
机构
[1] Guangdong Polytech Sci & Technol, Sch Robot, Zhuhai 519090, Peoples R China
[2] Harbin Univ Sci & Technol, Sch Elect & Elect Engn, Harbin 150000, Peoples R China
关键词
Multi-domain text classification; Multi-task learning; Recurrent Neural Netword(RNN); Convolutional Neural Network(CNN); NEURAL-NETWORK;
D O I
10.11999/JEIT200869
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In the text classification task, many texts in different domains are similarly expressed and have the characteristics of correlation, which can solve the problem of insufficient training data with labels. The text of different fields can be combined with the multi-task learning method, and the training accuracy and speed of the model can be improved. A Recurrent Convolution Multi-Task Learning (MTL-RC) model for text multi-classification is proposed, jointly modeling the text of multiple tasks, and taking advantage of multi-task learning, Recurrent Neural Network(RNN) and Convolutional Neural Network(CNN) models to obtain the correlation between multi-domain texts, long-term dependence of text. Local features of text are extracted. Rich experiments are carried out based on multi-domain text classification datasets, the Recurrent Convolution Multi-Task Learning(MTL-LC) proposed in this paper has an average accuracy of 90.1% for text classification in different fields, which is 6.5% higher than the single-task learning model STL-LC. Compared with mainstream multi-tasking learning models Full Shared Multi-Task Learning(FS-MTL), Adversarial Multi-Task Learninng(ASP-MTL), and Indirect Communciation for Multi-Task Learning(IC-MTL) have increased by 5.4%, 4%, and 2.8%, respectively.
引用
收藏
页码:2395 / 2403
页数:9
相关论文
共 18 条
  • [1] Long short-term memory
    Hochreiter, S
    Schmidhuber, J
    [J]. NEURAL COMPUTATION, 1997, 9 (08) : 1735 - 1780
  • [2] Blitzer J., 2007, ANN M ASS COMP LING, V45, P440
  • [3] Multitask learning
    Caruana, R
    [J]. MACHINE LEARNING, 1997, 28 (01) : 41 - 75
  • [4] Collobert R., 2008, P 25 ICML, V25, P160, DOI [DOI 10.1145/1390156.1390177, 10.1145/1390156.1390177]
  • [5] HUANG Xuanjing., 2016, 2016 C EMP METH NAT, P118
  • [6] A Convolutional Neural Network for Modelling Sentences
    Kalchbrenner, Nal
    Grefenstette, Edward
    Blunsom, Phil
    [J]. PROCEEDINGS OF THE 52ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1, 2014, : 655 - 665
  • [7] Kim Y, 2014, IEEE ASME INT C ADV, P1747, DOI 10.1109/AIM.2014.6878336
  • [8] LEE L., 2005, 43 ANN M ASS COMP LI, P115
  • [9] Multi-Domain Sentiment Classification with Classifier Combination
    Li, Shou-Shan
    Huang, Chu-Ren
    Zong, Cheng-Qing
    [J]. JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY, 2011, 26 (01) : 25 - 33
  • [10] Liu PF, 2019, AAAI CONF ARTIF INTE, P4360