Overlap in meaning is a stronger predictor of semantic activation in GPT-3 than in humans

被引:14
作者
Digutsch, Jan [1 ,3 ]
Kosinski, Michal [2 ]
机构
[1] Tech Univ Dortmund, Leibniz Res Ctr Working Environm & Human Factors, Dortmund, Germany
[2] Stanford Univ, Stanford, CA 94305 USA
[3] Univ St Gallen, Inst Behav Sci & Technol, St Gallen, Switzerland
关键词
LEXICAL DECISION; ASSOCIATION; MODELS;
D O I
10.1038/s41598-023-32248-6
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Modern large language models generate texts that are virtually indistinguishable from those written by humans and achieve near-human performance in comprehension and reasoning tests. Yet, their complexity makes it difficult to explain and predict their functioning. We examined a state-of-the-art language model (GPT-3) using lexical decision tasks widely used to study the structure of semantic memory in humans. The results of four analyses showed that GPT-3's patterns of semantic activation are broadly similar to those observed in humans, showing significantly higher semantic activation in related (e.g., "lime-lemon") word pairs than in other-related (e.g., "sour-lemon") or unrelated (e.g., "tourist-lemon") word pairs. However, there are also significant differences between GPT-3 and humans. GPT-3's semantic activation is better predicted by similarity in words' meaning (i.e., semantic similarity) rather than their co-occurrence in the language (i.e., associative similarity). This suggests that GPT-3's semantic network is organized around word meaning rather than their co-occurrence in text.
引用
收藏
页数:7
相关论文
共 35 条
  • [1] Aher G, 2022, Arxiv, DOI arXiv:2208.10264
  • [2] Brown TB, 2020, Arxiv, DOI [arXiv:2005.14165, DOI 10.48550/ARXIV.2005.14165]
  • [3] Baroni M, 2014, PROCEEDINGS OF THE 52ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1, P238
  • [4] SEMANTIC CONTEXT EFFECTS IN VISUAL WORD RECOGNITION - AN ANALYSIS OF SEMANTIC STRATEGIES
    BECKER, CA
    [J]. MEMORY & COGNITION, 1980, 8 (06) : 493 - 512
  • [5] On the Dangers of Stochastic Parrots: Can Language Models Be Too Big?
    Bender, Emily M.
    Gebru, Timnit
    McMillan-Major, Angelina
    Shmitchell, Shmargaret
    [J]. PROCEEDINGS OF THE 2021 ACM CONFERENCE ON FAIRNESS, ACCOUNTABILITY, AND TRANSPARENCY, FACCT 2021, 2021, : 610 - 623
  • [6] Benjamin AriS., 2019, 2019 Conference on Cognitive Computational Neuroscience, P585, DOI 10.32470/CCN.2019.1299-0
  • [7] Binz M., 2022, USING COGNITIVE PSYC, DOI [10.31234/osf.io/6dfgk, DOI 10.31234/OSF.IO/6DFGK]
  • [8] Cohen J., 1988, STAT POWER ANAL BEHA
  • [9] SPREADING ACTIVATION THEORY OF SEMANTIC PROCESSING
    COLLINS, AM
    LOFTUS, EF
    [J]. PSYCHOLOGICAL REVIEW, 1975, 82 (06) : 407 - 428
  • [10] Dasgupta S., P 60 ANN M ASS COMP, V1, P2263, DOI [10.18653/v1/2022.acl-long.161, DOI 10.18653/V1/2022.ACL-LONG.161]