Transfer Learning for Morphological Tagging in Russian

被引:0
作者
Andrianov, Ivan [1 ]
Mayorov, Vladimir [1 ]
机构
[1] Russian Acad Sci, Ivannikov Inst Syst Programming, Moscow, Russia
来源
2017 IVANNIKOV ISPRAS OPEN CONFERENCE (ISPRAS) | 2017年
关键词
morphological tagging; part-of-speech tagging; transfer learning; artificial neural networks;
D O I
10.1109/ISPRAS.2017.00017
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
This paper is devoted to morphological tagging task for Russian. There are multiple corpora created for the task during the years of research. Unfortunately, these corpora often have incompatible annotation guidelines and tag sets. This makes it difficult to utilize more than one corpus for training machine learning methods. Several attempts to manually unify guidelines and tag sets took place in the past. The most recent one was MorphoRuEval-2017 evaluation where organizers decided to unify annotations of four corpora. In this paper we propose a morphological tagging method which is able to be trained on any set of corpora with arbitrary guidelines and tag sets without any manual mapping provided. The method is based on transfer learning technique successfully applied in multilingual and cross-domain research. Firstly we establish a couple of solid neural network-based morphological tagging baselines. They both employ bidirectional LSTM networks but utilize different word embeddings. Comparative study on multiple corpora shows that fasttext model clearly outperforms word2vec one. Further we extend these baselines with transfer learning application. We prove transfer learning-based method effectiveness by performing two series of experiments. In the first series we exploit three wellknown corpora with incompatible tag sets: RNC, SynTagRus (original and Universal Dependencies versions). In the second series we utilize four corpora from MorphoRuEval-2017 evaluation with tag sets preliminary unified by organizers. In both cases transfer learning application improves performance of the method while training on more corpora. In the second case it even surpasses traditional joint training.
引用
收藏
页码:58 / 63
页数:6
相关论文
共 27 条
[1]  
[Anonymous], 2016, T ASSOC COMPUT LING, DOI DOI 10.1162/TACL_A_00051
[2]  
[Anonymous], 2016, P 30 AAAI C ART INT
[3]  
[Anonymous], 2014, P 18 C EMP METH NAT, DOI DOI 10.3115/V1/D14-1082
[4]  
[Anonymous], 2015, P EMNLP
[5]  
Apresjan J., 2006, Proceedings of LREC, P1378
[6]  
Brants T, 2000, 6TH APPLIED NATURAL LANGUAGE PROCESSING CONFERENCE/1ST MEETING OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, PROCEEDINGS OF THE CONFERENCE AND PROCEEDINGS OF THE ANLP-NAACL 2000 STUDENT RESEARCH WORKSHOP, P224
[7]  
BRILL E, 1992, SPEECH AND NATURAL LANGUAGE, P112
[8]  
Collobert R, 2011, J MACH LEARN RES, V12, P2493
[9]  
dos Santos CN, 2014, PR MACH LEARN RES, V32, P1818
[10]   Framewise phoneme classification with bidirectional LSTM and other neural network architectures [J].
Graves, A ;
Schmidhuber, J .
NEURAL NETWORKS, 2005, 18 (5-6) :602-610