Placing language in an integrated understanding system: Next steps toward human-level performance in neural language models

被引:55
作者
McClelland, James L. [1 ,2 ]
Hill, Felix [2 ]
Rudolph, Maja [3 ]
Baldridge, Jason [4 ]
Schutze, Hinrich [5 ]
机构
[1] Stanford Univ, Dept Psychol, Stanford, CA 94305 USA
[2] DeepMind, London N1C 4AG, England
[3] Bosch Ctr Artificial Intelligence, D-71272 Renningen, Germany
[4] Google Res, Austin, TX 78701 USA
[5] Ludwig Maximilian Univ Munich, Ctr Informat & Language Proc, D-80538 Munich, Germany
基金
欧洲研究理事会;
关键词
natural language understanding; deep learning; situation models; cognitive neuroscience; artificial intelligence; COMPLEMENTARY LEARNING-SYSTEMS; INTERACTIVE ACTIVATION; SEMANTIC KNOWLEDGE; MEMORY; PERCEPTION; REPRESENTATION; PRINCIPLES;
D O I
10.1073/pnas.1910416117
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Language is crucial for human intelligence, but what exactly is its role? We take language to be a part of a system for understanding and communicating about situations. In humans, these abilities emerge gradually from experience and depend on domain-general principles of biological neural networks: connection-based learning, distributed representation, and context-sensitive, mutual constraint satisfaction-based processing. Current artificial language processing systems rely on the same domain general principles, embodied in artificial neural networks. Indeed, recent progress in this field depends on query-based attention, which extends the ability of these systems to exploit context and has contributed to remarkable breakthroughs. Nevertheless, most current models focus exclusively on language-internal tasks, limiting their ability to perform tasks that depend on understanding situations. These systems also lack memory for the contents of prior situations outside of a fixed contextual span. We describe the organization of the brain's distributed understanding system, which includes a fast learning system that addresses the memory problem. We sketch a framework for future models of understanding drawing equally on cognitive neuroscience and artificial intelligence and exploiting query-based attention. We highlight relevant current directions and consider further developments needed to fully capture human-level language understanding in a computational system.
引用
收藏
页码:25966 / 25974
页数:9
相关论文
共 80 条
[11]   SENTENCE MEMORY - CONSTRUCTIVE VERSUS INTERPRETIVE APPROACH [J].
BRANSFORD, JD ;
BARCLAY, JR ;
FRANKS, JJ .
COGNITIVE PSYCHOLOGY, 1972, 3 (02) :193-209
[12]   CONTEXTUAL PREREQUISITES FOR UNDERSTANDING - SOME INVESTIGATIONS OF COMPREHENSION AND RECALL [J].
BRANSFORD, JD ;
JOHNSON, MK .
JOURNAL OF VERBAL LEARNING AND VERBAL BEHAVIOR, 1972, 11 (06) :717-726
[13]  
Brown TB, 2020, ADV NEUR IN, V33
[14]   Alternatives to the combinatorial paradigm of linguistic theory based on domain general principles of human cognition [J].
Bybee, J ;
Mcclelland, JL .
LINGUISTIC REVIEW, 2005, 22 (2-4) :381-410
[15]  
Chaplot DS, 2018, AAAI CONF ARTIF INTE, P2819
[16]  
Chomsky Noam., 1971, Semantics: an interdisciplinary reader, P183
[17]  
Collobert R, 2011, J MACH LEARN RES, V12, P2493
[18]  
Crain S., 1985, NATURAL LANGUAGE PAR, P320
[19]   Question Answering on Knowledge Bases and Text using Universal Schema and Memory Networks [J].
Das, Rajarshi ;
Zaheer, Manzil ;
Reddy, Siva ;
McCallum, Andrew .
PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 2, 2017, :358-365
[20]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171