Do large language models "understand" their knowledge?

被引:0
|
作者
Venkatasubramanian, Venkat [1 ]
机构
[1] Columbia Univ, Dept Chem Engn, Complex Resilient Intelligent Syst Lab, New York, NY 10027 USA
关键词
Knowledge representation; LLM; Industrial revolution 4.0; LKM; Transformers; PROCESS FAULT-DETECTION; QUANTITATIVE MODEL; PART I; FRAMEWORK; DESIGN; SYSTEM;
D O I
10.1002/aic.18661
中图分类号
TQ [化学工业];
学科分类号
0817 ;
摘要
Large language models (LLMs) are often criticized for lacking true "understanding" and the ability to "reason" with their knowledge, being seen merely as autocomplete engines. I suggest that this assessment might be missing a nuanced insight. LLMs do develop a kind of empirical "understanding" that is "geometry"-like, which is adequate for many applications. However, this "geometric" understanding, built from incomplete and noisy data, makes them unreliable, difficult to generalize, and lacking in inference capabilities and explanations. To overcome these limitations, LLMs should be integrated with an "algebraic" representation of knowledge that includes symbolic AI elements used in expert systems. This integration aims to create large knowledge models (LKMs) grounded in first principles that can reason and explain, mimicking human expert capabilities. Furthermore, we need a conceptual breakthrough, such as the transformation from Newtonian mechanics to statistical mechanics, to create a new science of LLMs.
引用
收藏
页数:10
相关论文
共 50 条
  • [31] Enrich Humanoids With Large Language Models (LLM)
    Antikatzidis, Angelos
    Feidakis, Michalis
    Marathaki, Konstantina
    Toumanidis, Lazaros
    Nikolaou, Grigoris
    Patrikakis, Charalampos Z.
    2024 IEEE GLOBAL ENGINEERING EDUCATION CONFERENCE, EDUCON 2024, 2024,
  • [32] Level Generation Through Large Language Models
    Todd, Graham
    Earle, Sam
    Nasir, Muhammad Umair
    Green, Michael Cerny
    Togelius, Julian
    PROCEEDINGS OF THE 18TH INTERNATIONAL CONFERENCE ON THE FOUNDATIONS OF DIGITAL GAMES, FDG 2023, 2023,
  • [33] Transformers and large language models in healthcare: A review
    Nerella, Subhash
    Bandyopadhyay, Sabyasachi
    Zhang, Jiaqing
    Contreras, Miguel
    Siegel, Scott
    Bumin, Aysegul
    Silva, Brandon
    Sena, Jessica
    Shickel, Benjamin
    Bihorac, Azra
    Khezeli, Kia
    Rashidi, Parisa
    ARTIFICIAL INTELLIGENCE IN MEDICINE, 2024, 154
  • [34] TOWARDS A CONVERSATIONAL ETHICS OF LARGE LANGUAGE MODELS
    Kempt, Hendrik
    Lavie, Alon
    Nagel, Saskia K.
    AMERICAN PHILOSOPHICAL QUARTERLY, 2024, 61 (04) : 339 - 354
  • [35] Assessing the Strengths and Weaknesses of Large Language Models
    Shalom Lappin
    Journal of Logic, Language and Information, 2024, 33 : 9 - 20
  • [36] Assessing the Strengths and Weaknesses of Large Language Models
    Lappin, Shalom
    JOURNAL OF LOGIC LANGUAGE AND INFORMATION, 2024, 33 (01) : 9 - 20
  • [37] Large language models and agricultural extension services
    Tzachor, A.
    Devare, M.
    Richards, C.
    Pypers, P.
    Ghosh, A.
    Koo, J.
    Johal, S.
    King, B.
    NATURE FOOD, 2023, 4 (11): : 941 - 948
  • [38] Large Language Models and the Future of Organization Theory
    Cornelissen, Joep
    Hollerer, Markus A.
    Boxenbaum, Eva
    Faraj, Samer
    Gehman, Joel
    ORGANIZATION THEORY, 2024, 5 (01):
  • [39] LARGE LANGUAGE MODELS (LLMS) AND CHATGPT FOR BIOMEDICINE
    Arighi, Cecilia
    Brenner, Steven
    Lu, Zhiyong
    BIOCOMPUTING 2024, PSB 2024, 2024, : 641 - 644
  • [40] Large Language Models in Gastroenterology: Systematic Review
    Gong, Eun Jeong
    Bang, Chang Seok
    Lee, Jae Jun
    Park, Jonghyung
    Kim, Eunsil
    Kim, Subeen
    Kimm, Minjae
    Choi, Seoung-Ho
    JOURNAL OF MEDICAL INTERNET RESEARCH, 2024, 26