Constraint satisfaction in large language models

被引:1
作者
Jacobs, Cassandra L. [1 ,3 ]
MacDonald, Maryellen C. [2 ]
机构
[1] SUNY Buffalo, Dept Linguist, Buffalo, NY USA
[2] Univ Wisconsin Madison, Dept Psychol, Madison, WI USA
[3] Univ Buffalo, Dept Linguist, Buffalo, NY 14260 USA
基金
美国国家科学基金会;
关键词
Language comprehension; constraint satisfaction; ambiguity; connectionism; large language models; WORD RECOGNITION; LEXICAL ACCESS; EYE-MOVEMENTS; AMBIGUITY; INFORMATION; CONTEXT; RESOLUTION; FIT;
D O I
10.1080/23273798.2024.2364339
中图分类号
R36 [病理学]; R76 [耳鼻咽喉科学];
学科分类号
100104 ; 100213 ;
摘要
Constraint satisfaction theories were prominent in the late 20th century and emphasized continuous, rich interaction between many sources of information in a linguistic signal unfolding over time. A major challenge was rigorously capturing these highly interactive comprehension processes and yielding explicit predictions, because the important constraints were numerous and changed in prominence from one context to the next. Connectionist models were conceptually well-suited to this, but researchers had insufficient computing power and lacked sufficiently large corpora to bring these models to bear. These limitations no longer hold, and large language models (LLMs) offer an opportunity to test constraint satisfaction ideas about human language comprehension. We consider how LLMs can be applied to study interactive processes with lexical ambiguity resolution as a test case. We argue that further study of LLMs can advance theories of constraint satisfaction, though gaps remain in our understanding of how people and LLMs combine linguistic information.
引用
收藏
页码:1231 / 1248
页数:18
相关论文
共 117 条
[1]  
Almeida R. G. D., 2018, CONCEPTS MODULES LAN
[2]   INTERACTION WITH CONTEXT DURING HUMAN SENTENCE PROCESSING [J].
ALTMANN, G ;
STEEDMAN, M .
COGNITION, 1988, 30 (03) :191-238
[3]   Ambiguity in sentence processing [J].
Altmann, GTM .
TRENDS IN COGNITIVE SCIENCES, 1998, 2 (04) :146-152
[4]  
[Anonymous], 1987, Empirical Model-Building and Response Surfaces
[5]   Predictive Coding or Just Feature Discovery? An Alternative Account of Why Language Models Fit Brain Data [J].
Antonello, Richard ;
Huth, Alexander .
NEUROBIOLOGY OF LANGUAGE, 2024, 5 (01) :64-79
[6]  
Asher N., 2023, ARXIV
[7]   Comprehension without segmentation: a proof of concept with naive discriminative learning [J].
Baayen, R. Harald ;
Shaoul, Cyrus ;
Willits, Jon ;
Ramscar, Michael .
LANGUAGE COGNITION AND NEUROSCIENCE, 2016, 31 (01) :106-128
[8]  
Bahdanau D, 2016, Arxiv, DOI arXiv:1409.0473
[9]  
Bianchi B., 2020, J VISION, V20, P1308, DOI [https://doi.org/10.1167/jov.20.11.1308, DOI 10.1167/JOV.20.11.1308]
[10]  
Bisk Y, 2020, PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), P8718