WINODICT: Probing language models for in-context word acquisition

被引:0
|
作者
Eisenschlos, Julian Martin [1 ]
Cole, Jeremy R. [1 ]
Liu, Fangyu [2 ]
Cohen, William W. [1 ]
机构
[1] Google Res, Mountain View, CA 94043 USA
[2] Univ Cambridge, Cambridge, England
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We introduce a new in-context learning paradigm to measure Large Language Models ' (LLMs) ability to learn novel words during inference. In particular, we rewrite Winogradstyle co-reference resolution problems by replacing the key concept word with a synthetic but plausible word that the model must understand to complete the task. Solving this task requires the model to make use of the dictionary definition of the new word given in the prompt. This benchmark addresses word acquisition, one important aspect of the diachronic degradation known to afflict LLMs. As LLMs are frozen in time at the moment they are trained, they are normally unable to reflect the way language changes over time. We show that the accuracy of LLMs compared to the original Winograd tasks decreases radically in our benchmark, thus identifying a limitation of current models and providing a benchmark to measure future improvements in LLMs ability to do in-context learning.
引用
收藏
页码:94 / 102
页数:9
相关论文
共 50 条
  • [41] Stabilized In-Context Learning with Pre-trained Language Models for Few Shot Dialogue State Tracking
    Chen, Derek
    Qian, Kun
    Yu, Zhou
    17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 1551 - 1564
  • [42] Cultural Understanding Using In-context Learning and Masked Language Modeling
    Qian, Ming
    Newton, Charles
    Qian, Davis
    HCI INTERNATIONAL 2021 - LATE BREAKING PAPERS: MULTIMODALITY, EXTENDED REALITY, AND ARTIFICIAL INTELLIGENCE, 2021, 13095 : 500 - 508
  • [43] THE ACQUISITION OF OBJECT NAMES IN CHILDREN WITH SPECIFIC LANGUAGE IMPAIRMENT - ACTION CONTEXT AND WORD EXTENSION
    SCHWARTZ, RG
    LEONARD, LB
    MESSICK, C
    CHAPMAN, K
    APPLIED PSYCHOLINGUISTICS, 1987, 8 (03) : 233 - 244
  • [44] In-context language control with production tasks in bilinguals: An fMRI study
    Zhang, Yong
    Huang, Peiyu
    Song, Zhe
    Fang, Liang
    Shen, Tong
    Li, Yan
    Gong, Qiyong
    Xie, Peng
    BRAIN RESEARCH, 2014, 1585 : 131 - 140
  • [45] Show Exemplars and Tell Me What You See: In-Context Learning with Frozen Large Language Models for TextVQA
    Zhang, Yan
    Zeng, Gangyan
    Shen, Huawen
    Ma, Can
    Zhou, Yu
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2024, PT VII, 2025, 15037 : 231 - 245
  • [46] Why Can GPT Learn In-Context? Language Models Implicitly Perform Gradient Descent as Meta-Optimizers
    Dai, Damai
    Sun, Yutao
    Dong, Li
    Hao, Yaru
    Ma, Shuming
    Sui, Zhifang
    Wei, Furu
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 4005 - 4019
  • [47] Towards Intent-based Configuration for Network Function Virtualization using In-context Learning in Large Language Models
    Nguyen Van Tu
    Yoo, Jae-Hyoung
    Hong, James Won-Ki
    PROCEEDINGS OF 2024 IEEE/IFIP NETWORK OPERATIONS AND MANAGEMENT SYMPOSIUM, NOMS 2024, 2024,
  • [48] Language Models can Exploit Cross-Task In-context Learning for Data-Scarce Novel Tasks
    Chatterjee, Anwoy
    Tanwar, Eshaan
    Dutta, Subhabrata
    Chakraborty, Tanmoy
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 11568 - 11587
  • [49] Linear Transformers with Learnable Kernel Functions are Better In-Context Models
    Aksenov, Yaroslav
    Balagansky, Nikita
    Vaina, Sofia Maria Lo Cicero
    Shaposhnikov, Boris
    Gorbatov, Alexey
    Gavrilov, Daniil
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 9584 - 9597
  • [50] Data Augmentation with In-Context Learning and Comparative Evaluation in Math Word Problem Solving
    Yigit G.
    Amasyali M.F.
    SN Computer Science, 5 (5)