WINODICT: Probing language models for in-context word acquisition

被引:0
|
作者
Eisenschlos, Julian Martin [1 ]
Cole, Jeremy R. [1 ]
Liu, Fangyu [2 ]
Cohen, William W. [1 ]
机构
[1] Google Res, Mountain View, CA 94043 USA
[2] Univ Cambridge, Cambridge, England
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We introduce a new in-context learning paradigm to measure Large Language Models ' (LLMs) ability to learn novel words during inference. In particular, we rewrite Winogradstyle co-reference resolution problems by replacing the key concept word with a synthetic but plausible word that the model must understand to complete the task. Solving this task requires the model to make use of the dictionary definition of the new word given in the prompt. This benchmark addresses word acquisition, one important aspect of the diachronic degradation known to afflict LLMs. As LLMs are frozen in time at the moment they are trained, they are normally unable to reflect the way language changes over time. We show that the accuracy of LLMs compared to the original Winograd tasks decreases radically in our benchmark, thus identifying a limitation of current models and providing a benchmark to measure future improvements in LLMs ability to do in-context learning.
引用
收藏
页码:94 / 102
页数:9
相关论文
共 50 条
  • [21] Large Language Models Can be Lazy Learners: Analyze Shortcuts in In-Context Learning
    Tang, Ruixiang
    Kong, Dehan
    Huang, Longtao
    Xue, Hui
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 4645 - 4657
  • [22] Concept-aware Data Construction Improves In-context Learning of Language Models
    Stefanik, Michal
    Kadlcik, Marek
    Sojka, Petr
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 12335 - 12352
  • [23] In-Context Learning Unlocked for Diffusion Models
    Wang, Zhendong
    Jiang, Yifan
    Lu, Yadong
    Shen, Yelong
    He, Pengcheng
    Chen, Weizhu
    Wang, Zhangyang
    Zhou, Mingyuan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [24] Automatic smart contract comment generation via large language models and in-context learning
    Zhao, Junjie
    Chen, Xiang
    Yang, Guang
    Shen, Yiheng
    INFORMATION AND SOFTWARE TECHNOLOGY, 2024, 168
  • [25] In-context learning enables multimodal large language models to classify cancer pathology images
    Ferber, Dyke
    Woelflein, Georg
    Wiest, Isabella C.
    Ligero, Marta
    Sainath, Srividhya
    Ghaffari Laleh, Narmin
    El Nahhas, Omar S. M.
    Mueller-Franzes, Gustav
    Jaeger, Dirk
    Truhn, Daniel
    Kather, Jakob Nikolas
    NATURE COMMUNICATIONS, 2024, 15 (01)
  • [26] Query-focused Submodular Demonstration Selection for In-context Learning in Large Language Models
    Trust, Paul
    Minghim, Rosane
    2023 31ST IRISH CONFERENCE ON ARTIFICIAL INTELLIGENCE AND COGNITIVE SCIENCE, AICS, 2023,
  • [27] MAGNIFICO: Evaluating the In-Context Learning Ability of Large Language Models to Generalize to Novel Interpretations
    Patel, Arkil
    Bhattamishra, Satwik
    Reddy, Siva
    Bahdanau, Dzmitry
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 2167 - 2189
  • [28] ScoNe: Benchmarking Negation Reasoning in Language Models With Fine-Tuning and In-Context Learning
    She, Jingyuan Selena
    Potts, Christopher
    Bowman, Samuel R.
    Geiger, Atticus
    61ST CONFERENCE OF THE THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 2, 2023, : 1803 - 1821
  • [29] GPT-RE: In-context Learning for Relation Extraction using Large Language Models
    Wan, Zhen
    Cheng, Fei
    Mao, Zhuoyuan
    Liu, Qianying
    Song, Haiyue
    Li, Jiwei
    Kurohashi, Sadao
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 3534 - 3547
  • [30] Trained Transformers Learn Linear Models In-Context
    Zhang, Ruiqi
    Frei, Spencer
    Bartlett, Peter L.
    JOURNAL OF MACHINE LEARNING RESEARCH, 2024, 25