Constraint satisfaction in large language models

被引:0
作者
Jacobs, Cassandra L. [1 ,3 ]
MacDonald, Maryellen C. [2 ]
机构
[1] SUNY Buffalo, Dept Linguist, Buffalo, NY USA
[2] Univ Wisconsin Madison, Dept Psychol, Madison, WI USA
[3] Univ Buffalo, Dept Linguist, Buffalo, NY 14260 USA
基金
美国国家科学基金会;
关键词
Language comprehension; constraint satisfaction; ambiguity; connectionism; large language models; WORD RECOGNITION; LEXICAL ACCESS; EYE-MOVEMENTS; AMBIGUITY; INFORMATION; CONTEXT; RESOLUTION; FIT;
D O I
10.1080/23273798.2024.2364339
中图分类号
R36 [病理学]; R76 [耳鼻咽喉科学];
学科分类号
100104 ; 100213 ;
摘要
Constraint satisfaction theories were prominent in the late 20th century and emphasized continuous, rich interaction between many sources of information in a linguistic signal unfolding over time. A major challenge was rigorously capturing these highly interactive comprehension processes and yielding explicit predictions, because the important constraints were numerous and changed in prominence from one context to the next. Connectionist models were conceptually well-suited to this, but researchers had insufficient computing power and lacked sufficiently large corpora to bring these models to bear. These limitations no longer hold, and large language models (LLMs) offer an opportunity to test constraint satisfaction ideas about human language comprehension. We consider how LLMs can be applied to study interactive processes with lexical ambiguity resolution as a test case. We argue that further study of LLMs can advance theories of constraint satisfaction, though gaps remain in our understanding of how people and LLMs combine linguistic information.
引用
收藏
页码:1231 / 1248
页数:18
相关论文
共 117 条
  • [1] Almeida R. G. D., 2018, CONCEPTS MODULES LAN
  • [2] INTERACTION WITH CONTEXT DURING HUMAN SENTENCE PROCESSING
    ALTMANN, G
    STEEDMAN, M
    [J]. COGNITION, 1988, 30 (03) : 191 - 238
  • [3] Ambiguity in sentence processing
    Altmann, GTM
    [J]. TRENDS IN COGNITIVE SCIENCES, 1998, 2 (04) : 146 - 152
  • [4] Predictive Coding or Just Feature Discovery? An Alternative Account of Why Language Models Fit Brain Data
    Antonello, Richard
    Huth, Alexander
    [J]. NEUROBIOLOGY OF LANGUAGE, 2024, 5 (01): : 64 - 79
  • [5] Asher N., 2023, ARXIV
  • [6] Comprehension without segmentation: a proof of concept with naive discriminative learning
    Baayen, R. Harald
    Shaoul, Cyrus
    Willits, Jon
    Ramscar, Michael
    [J]. LANGUAGE COGNITION AND NEUROSCIENCE, 2016, 31 (01) : 106 - 128
  • [7] Bahdanau D, 2016, Arxiv, DOI arXiv:1409.0473
  • [8] Bianchi B., 2020, J VISION, V20, P1308, DOI [https://doi.org/10.1167/jov.20.11.1308, DOI 10.1167/JOV.20.11.1308]
  • [9] Bisk Y, 2020, PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), P8718
  • [10] What are large language models supposed to model?
    Blank, Idan A.
    [J]. TRENDS IN COGNITIVE SCIENCES, 2023, 27 (11) : 987 - 989