WINODICT: Probing language models for in-context word acquisition

被引:0
|
作者
Eisenschlos, Julian Martin [1 ]
Cole, Jeremy R. [1 ]
Liu, Fangyu [2 ]
Cohen, William W. [1 ]
机构
[1] Google Res, Mountain View, CA 94043 USA
[2] Univ Cambridge, Cambridge, England
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We introduce a new in-context learning paradigm to measure Large Language Models ' (LLMs) ability to learn novel words during inference. In particular, we rewrite Winogradstyle co-reference resolution problems by replacing the key concept word with a synthetic but plausible word that the model must understand to complete the task. Solving this task requires the model to make use of the dictionary definition of the new word given in the prompt. This benchmark addresses word acquisition, one important aspect of the diachronic degradation known to afflict LLMs. As LLMs are frozen in time at the moment they are trained, they are normally unable to reflect the way language changes over time. We show that the accuracy of LLMs compared to the original Winograd tasks decreases radically in our benchmark, thus identifying a limitation of current models and providing a benchmark to measure future improvements in LLMs ability to do in-context learning.
引用
收藏
页码:94 / 102
页数:9
相关论文
共 50 条
  • [31] From System Models to Class Models: An In-Context Learning Paradigm
    Forgione, Marco
    Pura, Filippo
    Piga, Dario
    IEEE CONTROL SYSTEMS LETTERS, 2023, 7 : 3513 - 3518
  • [32] Self-Augmented In-Context Learning for Unsupervised Word Translation
    Li, Yaoyiran
    Korhonen, Anna
    Vulic, Ivan
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 2: SHORT PAPERS, 2024, : 743 - 753
  • [33] Customizing Language Model Responses with Contrastive In-Context Learning
    Gao, Xiang
    Das, Kamalika
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 16, 2024, : 18039 - 18046
  • [34] Language Models as Context-sensitive Word Search Engines
    Wiegmann, Matti
    Voelske, Michael
    Stein, Benno
    Potthast, Martin
    PROCEEDINGS OF THE FIRST WORKSHOP ON INTELLIGENT AND INTERACTIVE WRITING ASSISTANTS (IN2WRITING 2022), 2022, : 39 - 45
  • [35] Distinguishing word identity and sequence context in DNA language models
    Sanabria, Melissa
    Hirsch, Jonas
    Poetsch, Anna R.
    BMC BIOINFORMATICS, 2024, 25 (01):
  • [36] Improving Grapheme-to-Phoneme Conversion through In-Context Knowledge Retrieval with Large Language Models
    Han, Dongrui
    Cui, Mingyu
    Kang, Jiawen
    Wu, Xixin
    Liu, Xunying
    Meng, Helen
    2024 IEEE 14TH INTERNATIONAL SYMPOSIUM ON CHINESE SPOKEN LANGUAGE PROCESSING, ISCSLP 2024, 2024, : 631 - 635
  • [37] SCALa: A blueprint for computational models of language acquisition in social context
    Tsuji, Sho
    Cristia, Alejandrina
    Dupoux, Emmanuel
    COGNITION, 2021, 213
  • [38] A word-order database for testing computational models of language acquisition
    Sakas, WG
    41ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, PROCEEDINGS OF THE CONFERENCE, 2003, : 415 - 422
  • [39] Structured State Space Models for In-Context Reinforcement Learning
    Lu, Chris
    Schroecker, Yannick
    Gu, Albert
    Parisotto, Emilio
    Foerster, Jakob
    Singh, Satinder
    Behbahani, Feryal
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [40] Meta-learning via Language Model In-context Tuning
    Chen, Yanda
    Zhong, Ruiqi
    Zha, Sheng
    Karypis, George
    He, He
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 719 - 730