Mention detection in Turkish coreference resolution

被引:0
作者
Demir, Seniz [1 ]
Akdag, Hanifi Ibrahim [1 ]
机构
[1] MEF Univ, Dept Comp Engn, Istanbul, Turkiye
关键词
Coreference resolution; mention detection; neural network; language model; Turkish;
D O I
10.55730/1300-0632.4095
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A crucial step in understanding natural language is detecting mentions that refer to real-world entities in a text and correctly identifying their boundaries. Mention detection is commonly considered a preprocessing step in coreference resolution which is shown to be helpful in several language processing applications such as machine translation and text summarization. Despite recent efforts on Turkish coreference resolution, no standalone neural solution to mention detection has been proposed yet. In this article, we present two models designed for detecting Turkish mentions by using feed-forward neural networks. Both models extract all spans up to a fixed length from input text as candidates and classify them as mentions or not mentions. The models differ in terms of how candidate text spans are represented. The first model represents a span by focusing on its first and last words, whereas the representation also covers the preceding and proceeding words of a span in the second model. Mention span representations are formed by using contextual embeddings, part-of-speech embeddings, and named-entity embeddings of words in interest where contextual embeddings are obtained from pretrained Turkish language models. In our evaluation studies, we not only assess the impact of mention representation strategies on system performance but also demonstrate the usability of different pretrained language models in resolution task. We argue that our work provides useful insights to the existing literature and the first step in understanding the effectiveness of neural architectures in Turkish mention detection.
引用
收藏
页码:682 / 697
页数:17
相关论文
共 57 条
  • [1] Aloraini Abdulrahman., 2020, 3 WORKSH COMP MOD RE, P99
  • [2] ArsIan TP, 2023, 17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, P14
  • [3] Barua A, 2017, Semantic Scholar
  • [4] Investigating Query Expansion and Coreference Resolution in Question Answering on BERT
    Bhattacharjee, Santanu
    Haque, Rejwanul
    Wenniger, Gideon Maillette De Buy
    Way, Andy
    [J]. NATURAL LANGUAGE PROCESSING AND INFORMATION SYSTEMS (NLDB 2020), 2020, 12089 : 47 - 59
  • [5] Bis D, 2021, 2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), P5117
  • [6] Clark K., 2016, ARXIV160908667
  • [7] Coreference Resolution: Toward End-to-End and Cross-Lingual Systems
    Cruz, Andre Ferreira
    Rocha, Gil
    Cardoso, Henrique Lopes
    [J]. INFORMATION, 2020, 11 (02)
  • [8] Daum'e Hal., 2005, P C HUMAN LANGUAGE T, P97
  • [9] Demir S, 2022, INT C INF COMP SCI, P39
  • [10] Demir S., 2023, Journal of Intelligent Systems: Theory and Applications, V6, P85, DOI DOI 10.38016/JISTA.1225097