Understanding models understanding language

被引:7
|
作者
Sogaard, Anders [1 ,2 ]
机构
[1] Univ Copenhagen, Pioneer Ctr Artificial Intelligence, Dept Comp Sci, Lyngbyvej 2, DK-2100 Copenhagen, Denmark
[2] Univ Copenhagen, Dept Philosophy, Lyngbyvej 2, DK-2100 Copenhagen, Denmark
关键词
Artificial intelligence; Language; Mind; BRAINS;
D O I
10.1007/s11229-022-03931-4
中图分类号
N09 [自然科学史]; B [哲学、宗教];
学科分类号
01 ; 0101 ; 010108 ; 060207 ; 060305 ; 0712 ;
摘要
Landgrebe and Smith (Synthese 198(March):2061-2081, 2021) present an unflattering diagnosis of recent advances in what they call language-centric artificial intelligence-perhaps more widely known as natural language processing: The models that are currently employed do not have sufficient expressivity, will not generalize, and are fundamentally unable to induce linguistic semantics, they say. The diagnosis is mainly derived from an analysis of the widely used Transformer architecture. Here I address a number of misunderstandings in their analysis, and present what I take to be a more adequate analysis of the ability of Transformer models to learn natural language semantics. To avoid confusion, I distinguish between inferential and referential semantics. Landgrebe and Smith (2021)'s analysis of the Transformer architecture's expressivity and generalization concerns inferential semantics. This part of their diagnosis is shown to rely on misunderstandings of technical properties of Transformers. Landgrebe and Smith (2021) also claim that referential semantics is unobtainable for Transformer models. In response, I present a non-technical discussion of techniques for grounding Transformer models, giving them referential semantics, even in the absence of supervision. I also present a simple thought experiment to highlight the mechanisms that would lead to referential semantics, and discuss in what sense models that are grounded in this way, can be said to understand language. Finally, I discuss the approach Landgrebe and Smith (2021) advocate for, namely manual specification of formal grammars that associate linguistic expressions with logical form.
引用
收藏
页数:16
相关论文
共 50 条
  • [1] Understanding models understanding language
    Anders Søgaard
    Synthese, 200
  • [2] The Journey of Language Models in Understanding Natural Language
    Liu, Yuanrui
    Zhou, Jingping
    Sang, Guobiao
    Huang, Ruilong
    Zhao, Xinzhe
    Fang, Jintao
    Wang, Tiexin
    Li, Bohan
    WEB INFORMATION SYSTEMS AND APPLICATIONS, WISA 2024, 2024, 14883 : 331 - 363
  • [3] Meaning and understanding in large language models
    Havlik, Vladimir
    SYNTHESE, 2024, 205 (01)
  • [4] Understanding, Design, Models, Dialogue: The Orienting Role of Language
    Richards, Larry
    SHE JI-THE JOURNAL OF DESIGN ECONOMICS AND INNOVATION, 2019, 5 (04) : 369 - 372
  • [5] Understanding natural language: Potential application of large language models to ophthalmology
    Yang, Zefeng
    Wang, Deming
    Zhou, Fengqi
    Song, Diping
    Zhang, Yinhang
    Jiang, Jiaxuan
    Kong, Kangjie
    Liu, Xiaoyi
    Qiao, Yu
    Chang, Robert T.
    Han, Ying
    Li, Fei
    Tham, Clement C.
    Zhang, Xiulan
    ASIA-PACIFIC JOURNAL OF OPHTHALMOLOGY, 2024, 13 (04):
  • [6] fine-grained comparison of pragmatic language understanding in humans and language models
    Hu, Jennifer
    Floyd, Sammy
    Jouravlev, Olessia
    Fedorenko, Evelina
    Gibson, Edward
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 4194 - 4213
  • [7] The debate over understanding in AI?s large language models
    Mitchell, Melanie
    Krakauer, David C.
    PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2023, 120 (13)
  • [8] Natural Language Understanding: Methodological Conceptualization
    Shymko, Vitalii
    PSYCHOLINGUISTICS, 2019, 25 (01): : 431 - 443
  • [9] Language and understanding (1970)
    Gadamer, HG
    THEORY CULTURE & SOCIETY, 2006, 23 (01) : 13 - +
  • [10] Understanding Sarcoidosis Using Large Language Models and Social Media Data
    Xi, Nan Miles
    Ji, Hong-Long
    Wang, Lin
    JOURNAL OF HEALTHCARE INFORMATICS RESEARCH, 2024,