Large language models are better than theoretical linguists at theoretical linguistics

被引:0
作者
Ambridge, Ben [1 ,2 ]
Blything, Liam [1 ,2 ]
机构
[1] Univ Manchester, Manchester, England
[2] ESRCInternat Ctr Language & Commun Dev LuCiD, Manchester, England
基金
英国经济与社会研究理事会; 欧洲研究理事会;
关键词
large language models; causatives; grammaticality judgments; VERB SEMANTICS; ENTRENCHMENT; RETREAT; CONSTRAINTS; PREEMPTION; ERRORS; ROLES; SAY;
D O I
10.1515/tl-2024-2002
中图分类号
H0 [语言学];
学科分类号
030303 ; 0501 ; 050102 ;
摘要
Large language models are better than theoretical linguists at theoretical linguistics, at least in the domain of verb argument structure; explaining why (for example), we can say both The ball rolled and Someone rolled the ball, but not both The man laughed and *Someone laughed the man. Verbal accounts of this phenomenon either do not make precise quantitative predictions at all, or do so only with the help of ancillary assumptions and by-hand data processing. Large language models, on the other hand (taking text-davinci-002 as an example), predict human acceptability ratings for these types of sentences with correlations of around r = 0.9, and themselves constitute theories of language acquisition and representation; theories that instantiate exemplar-, input- and construction-based approaches, though only very loosely. Indeed, large language models succeed where these verbal (i.e., non-computational) linguistic theories fail, precisely because the latter insist - in the service of intuitive interpretability - on simple yet empirically inadequate (over)generalizations.
引用
收藏
页码:33 / 48
页数:16
相关论文
共 50 条
  • [31] Manner implicatures in large language models
    Cong, Yan
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [32] A survey on LoRA of large language models
    Mao, Yuren
    Ge, Yuhang
    Fan, Yijiang
    Xu, Wenyi
    Mi, Yu
    Hu, Zhonghao
    Gao, Yunjun
    FRONTIERS OF COMPUTER SCIENCE, 2025, 19 (07)
  • [33] Meaning and understanding in large language models
    Havlik, Vladimir
    SYNTHESE, 2024, 205 (01)
  • [34] A Survey on Evaluation of Large Language Models
    Chang, Yupeng
    Wang, Xu
    Wang, Jindong
    Wu, Yuan
    Yang, Linyi
    Zhu, Kaijie
    Chen, Hao
    Yi, Xiaoyuan
    Wang, Cunxiang
    Wang, Yidong
    Ye, Wei
    Zhang, Yue
    Chang, Yi
    Yu, Philip S.
    Yang, Qiang
    Xie, Xing
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2024, 15 (03)
  • [35] Emotional intelligence of Large Language Models
    Wang, Xuena
    Li, Xueting
    Yin, Zi
    Wu, Yue
    Liu, Jia
    JOURNAL OF PACIFIC RIM PSYCHOLOGY, 2023, 17
  • [36] Technical foundations of large language models
    Bluethgen, Christian
    RADIOLOGIE, 2025, : 227 - 234
  • [37] Large language models and their applications in bioinformatics
    Sarumi, Oluwafemi A.
    Heider, Dominik
    COMPUTATIONAL AND STRUCTURAL BIOTECHNOLOGY JOURNAL, 2024, 23 : 3498 - 3505
  • [38] On the Question of Authorship in Large Language Models
    Soos, Carlin
    Haroutunian, Levon
    KNOWLEDGE ORGANIZATION, 2024, 51 (02): : 83 - 95
  • [39] Large language models and linguistic intentionality
    Grindrod, Jumbly
    SYNTHESE, 2024, 204 (02)
  • [40] Process Modeling with Large Language Models
    Kourani, Humam
    Berti, Alessandro
    Schuster, Daniel
    van der Aalst, Wil M. P.
    ENTERPRISE, BUSINESS-PROCESS AND INFORMATION SYSTEMS MODELING, BPMDS 2024, EMMSAD 2024, 2024, 511 : 229 - 244