Large language models are better than theoretical linguists at theoretical linguistics

被引:0
作者
Ambridge, Ben [1 ,2 ]
Blything, Liam [1 ,2 ]
机构
[1] Univ Manchester, Manchester, England
[2] ESRCInternat Ctr Language & Commun Dev LuCiD, Manchester, England
基金
英国经济与社会研究理事会; 欧洲研究理事会;
关键词
large language models; causatives; grammaticality judgments; VERB SEMANTICS; ENTRENCHMENT; RETREAT; CONSTRAINTS; PREEMPTION; ERRORS; ROLES; SAY;
D O I
10.1515/tl-2024-2002
中图分类号
H0 [语言学];
学科分类号
030303 ; 0501 ; 050102 ;
摘要
Large language models are better than theoretical linguists at theoretical linguistics, at least in the domain of verb argument structure; explaining why (for example), we can say both The ball rolled and Someone rolled the ball, but not both The man laughed and *Someone laughed the man. Verbal accounts of this phenomenon either do not make precise quantitative predictions at all, or do so only with the help of ancillary assumptions and by-hand data processing. Large language models, on the other hand (taking text-davinci-002 as an example), predict human acceptability ratings for these types of sentences with correlations of around r = 0.9, and themselves constitute theories of language acquisition and representation; theories that instantiate exemplar-, input- and construction-based approaches, though only very loosely. Indeed, large language models succeed where these verbal (i.e., non-computational) linguistic theories fail, precisely because the latter insist - in the service of intuitive interpretability - on simple yet empirically inadequate (over)generalizations.
引用
收藏
页码:33 / 48
页数:16
相关论文
共 50 条
  • [21] Dual Adapter Tuning of Vision–Language Models Using Large Language Models
    Mohammad Reza Zarei
    Abbas Akkasi
    Majid Komeili
    International Journal of Computational Intelligence Systems, 18 (1)
  • [22] Industrial applications of large language models
    Mubashar Raza
    Zarmina Jahangir
    Muhammad Bilal Riaz
    Muhammad Jasim Saeed
    Muhammad Awais Sattar
    Scientific Reports, 15 (1)
  • [23] Large Language Models in Finance: A Survey
    Li, Yinheng
    Wang, Shaofei
    Ding, Han
    Chen, Hang
    PROCEEDINGS OF THE 4TH ACM INTERNATIONAL CONFERENCE ON AI IN FINANCE, ICAIF 2023, 2023, : 374 - 382
  • [24] Explainability for Large Language Models: A Survey
    Zhao, Haiyan
    Chen, Hanjie
    Yang, Fan
    Liu, Ninghao
    Deng, Huiqi
    Cai, Hengyi
    Wang, Shuaiqiang
    Yin, Dawei
    Du, Mengnan
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2024, 15 (02)
  • [25] On the attribution of confidence to large language models
    Keeling, Geoff
    Street, Winnie
    INQUIRY-AN INTERDISCIPLINARY JOURNAL OF PHILOSOPHY, 2025,
  • [26] Tutorial on Large Language Models for Recommendation
    Hua, Wenyue
    Li, Lei
    Xu, Shuyuan
    Chen, Li
    Zhang, Yongfeng
    PROCEEDINGS OF THE 17TH ACM CONFERENCE ON RECOMMENDER SYSTEMS, RECSYS 2023, 2023, : 1281 - 1283
  • [27] Large language models for chemistry robotics
    Naruki Yoshikawa
    Marta Skreta
    Kourosh Darvish
    Sebastian Arellano-Rubach
    Zhi Ji
    Lasse Bjørn Kristensen
    Andrew Zou Li
    Yuchi Zhao
    Haoping Xu
    Artur Kuramshin
    Alán Aspuru-Guzik
    Florian Shkurti
    Animesh Garg
    Autonomous Robots, 2023, 47 : 1057 - 1086
  • [28] Fusion Pruning for Large Language Models
    Jiang, Shixin
    Liu, Ming
    Qin, Bing
    2024 IEEE 14TH INTERNATIONAL SYMPOSIUM ON CHINESE SPOKEN LANGUAGE PROCESSING, ISCSLP 2024, 2024, : 349 - 352
  • [29] Risk communication and large language models
    Sledge, Daniel
    Thomas, Herschel F.
    RISK HAZARDS & CRISIS IN PUBLIC POLICY, 2024,
  • [30] On the Unexpected Abilities of Large Language Models
    Nolfi, Stefano
    ADAPTIVE BEHAVIOR, 2024, 32 (06) : 493 - 502