A Cluster Ranking Model for Full Anaphora Resolution

被引:0
作者
Yu, Juntao [1 ]
Uma, Alexandra [1 ]
Poesio, Massimo [1 ]
机构
[1] Queen Mary Univ London, London, England
来源
PROCEEDINGS OF THE 12TH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION (LREC 2020) | 2020年
关键词
Anaphora Resolution; Coreference; Cluster ranking model; Non-referring detection; Deep Neural Networks;
D O I
暂无
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Anaphora resolution (coreference) systems designed for the CONLL 2012 dataset typically cannot handle key aspects of the full anaphora resolution task such as the identification of singletons and of certain types of non-referring expressions (e.g., expletives), as these aspects are not annotated in that corpus. However, the recently released CRAC 2018 Shared Task and Phrase Detectives (PD) datasets can now be used for that purpose. In this paper, we introduce an architecture to simultaneously identify non-referring expressions (including expletives, predicative NPs, and other types) and build coreference chains, including singletons. Our cluster-ranking system uses an attention mechanism to determine the relative importance of the mentions in the same cluster. Additional classifiers are used to identify singletons and non-referring markables. Our contributions are as follows. First of all, we report the first result on the CRAC data using system mentions; our result is 5.8% better than the shared task baseline system, which used gold mentions. Our system also outperforms the best-reported system on PD by up to 5.3%. Second, we demonstrate that the availability of singleton clusters and non-referring expressions can lead to substantially improved performance on non-singleton clusters as well. Third, we show that despite our model not being designed specifically for the CONLL data, it achieves a very competitive result.
引用
收藏
页码:11 / 20
页数:10
相关论文
共 47 条
[1]  
[Anonymous], 2008, P 48 ANN M ASS COMPU, P10
[2]  
Ariel Mira, 1990, Accessing Noun-Phrase Antecedents
[3]  
Bahdanau D, 2016, Arxiv, DOI arXiv:1409.0473
[4]  
Bergsma S., 2011, Proceedings of DAARC, P12
[5]  
Björkelund A, 2014, PROCEEDINGS OF THE 52ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1, P47
[6]  
Boyd A., 2005, Proceedings of the ACL Workshop on Feature Engineering for Machine Learning in Natural Language Processing, P40
[7]  
Chen H, 2018, 2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), P172
[8]  
Clark K., 2016, P EMNLP
[9]  
Clark K., 2016, ASS COMPUTATIONAL LI
[10]  
Clark K, 2015, PROCEEDINGS OF THE 53RD ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 7TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 1, P1405