Cross-Lingual Universal Dependency Parsing Only From One Monolingual Treebank

被引:4
作者
Sun, Kailai [1 ]
Li, Zuchao [2 ]
Zhao, Hai [1 ]
机构
[1] Shanghai Jiao Tong Univ, AI Inst, Dept Comp Sci & Engn,MoE Key Lab Artificial Intell, Key Lab Shanghai Educ Commiss Intelligent Interact, Shanghai 200240, Peoples R China
[2] Wuhan Univ, Sch Comp Sci, Wuhan 430072, Peoples R China
基金
中国国家自然科学基金;
关键词
Training; Task analysis; Data models; Syntactics; Annotations; Transfer learning; Silver; Universal dependency parsing; few-shot parsing; zero-shot parsing; cross-lingual language processing; self-training;
D O I
10.1109/TPAMI.2023.3291388
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Syntactic parsing is a highly linguistic processing task whose parser requires training on treebanks from the expensive human annotation. As it is unlikely to obtain a treebank for every human language, in this work, we propose an effective cross-lingual UD parsing framework for transferring parser from only one source monolingual treebank to any other target languages without treebank available. To reach satisfactory parsing accuracy among quite different languages, we introduce two language modeling tasks into the training process of dependency parsing as multi-tasking. Assuming only unlabeled data from target languages plus the source treebank can be exploited together, we adopt a self-training strategy for further performance improvement in terms of our multi-task framework. Our proposed cross-lingual parsers are implemented for English, Chinese, and 29 UD treebanks. The empirical study shows that our cross-lingual parsers yield promising results for all target languages, approaching the parser performance which is trained in its own target treebank.
引用
收藏
页码:13393 / 13407
页数:15
相关论文
共 83 条
[1]   Unsupervised Domain Adaptation of Deep Networks for ToF Depth Refinement [J].
Agresti, Gianluca ;
Schaefer, Henrik ;
Sartor, Piergiorgio ;
Incesu, Yalcin ;
Zanuttigh, Pietro .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (12) :9195-9208
[2]  
Ahmad W. U., 2019, Proceedings of CoNLL, P372
[3]  
Ahmad WU, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P2440
[4]  
[Anonymous], 2016, Transactions of the Association for Computational Linguistics
[5]  
[Anonymous], 1993, Comput. Linguist, DOI DOI 10.21236/ADA273556
[6]  
Bahdanau D, 2016, Arxiv, DOI [arXiv:1409.0473, DOI 10.48550/ARXIV.1409.0473]
[7]  
Berthelot D, 2019, ADV NEUR IN, V32
[8]  
Bojanowski P, 2017, Transactions of the Association for Computational Linguistics, V5, P135, DOI [DOI 10.1162/TACL_A_00051, 10.1162/tacla00051]
[9]   Unsupervised Transfer Learning via Multi-Scale Convolutional Sparse Coding for Biomedical Applications [J].
Chang, Hang ;
Han, Ju ;
Zhong, Cheng ;
Snijders, Antoine M. ;
Mao, Jian-Hua .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (05) :1182-1194
[10]  
CHU YJ, 1965, SCI SINICA, V14, P1396