Language Model Priming for Cross-Lingual Event Extraction

被引:0
|
作者
Fincke, Steven [1 ]
Agarwal, Shantanu [1 ]
Miller, Scott [1 ]
Boschee, Elizabeth [1 ]
机构
[1] Univ Southern Calif, Inst Informat Sci, Los Angeles, CA 90007 USA
来源
THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE | 2022年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present a novel, language-agnostic approach to "priming" language models for the task of event extraction, providing particularly effective performance in low-resource and zero-shot cross-lingual settings. With priming, we augment the input to the transformer stack's language model differently depending on the question(s) being asked of the model at runtime. For instance, if the model is being asked to identify arguments for the trigger protested, we will provide that trigger as part of the input to the language model, allowing it to produce different representations for candidate arguments than when it is asked about arguments for the trigger arrest elsewhere in the same sentence. We show that by enabling the language model to better compensate for the deficits of sparse and noisy training data, our approach improves both trigger and argument detection and classification significantly over the state of the art in a zero-shot cross-lingual setting.
引用
收藏
页码:10627 / 10635
页数:9
相关论文
共 50 条
  • [41] DiffSLU: Knowledge Distillation Based Diffusion Model for Cross-Lingual Spoken Language Understanding
    Mao, Tianjun
    Zhang, Chenghong
    INTERSPEECH 2023, 2023, : 715 - 719
  • [42] A syntax-enhanced parameter generation network for multi-source cross-lingual event extraction
    Huang, Wenzhi
    Zhang, Junchi
    Tian, Tian
    Ji, Donghong
    KNOWLEDGE-BASED SYSTEMS, 2024, 292
  • [43] An analysis on language transfer of pre-trained language model with cross-lingual post-training
    Son, Suhyune
    Park, Chanjun
    Lee, Jungseob
    Shim, Midan
    Lee, Chanhee
    Jang, Yoonna
    Seo, Jaehyung
    Lim, Jungwoo
    Lim, Heuiseok
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 267
  • [44] END-to-END Cross-Lingual Spoken Language Understanding Model with Multilingual Pretraining
    Zhang, Xianwei
    He, Liang
    INTERSPEECH 2021, 2021, : 4728 - 4732
  • [45] END-TO-END CODE-SWITCHING TTS WITH CROSS-LINGUAL LANGUAGE MODEL
    Zhou, Xuehao
    Tian, Xiaohai
    Lee, Grandee
    Das, Rohan Kumar
    Li, Haizhou
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 7614 - 7618
  • [46] Allocating Large Vocabulary Capacity for Cross-lingual Language Model Pre-training
    Zheng, Bo
    Dong, Li
    Huang, Shaohan
    Singhal, Saksham
    Che, Wanxiang
    Liu, Ting
    Song, Xia
    Wei, Furu
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 3203 - 3215
  • [47] Zero-Shot Cross-Lingual Opinion Target Extraction
    Jebbara, Soufian
    Cimiano, Philipp
    2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, 2019, : 2486 - 2495
  • [48] Cross-Lingual Consistency of Factual Knowledge in Multilingual Language Models
    Qi, Jirui
    Fernandez, Raquel
    Bisazza, Arianna
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 10650 - 10666
  • [49] Cross-Lingual Lexico-Semantic Transfer in Language Learning
    Kochmar, Ekaterina
    Shutova, Ekaterina
    PROCEEDINGS OF THE 54TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1, 2016, : 974 - 983
  • [50] Steering Large Language Models for Cross-lingual Information Retrieval
    Guo, Ping
    Ren, Yubing
    Hu, Yue
    Cao, Yanan
    Li, Yunpeng
    Huang, Heyan
    PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 585 - 596