Placing language in an integrated understanding system: Next steps toward human-level performance in neural language models

被引:55
作者
McClelland, James L. [1 ,2 ]
Hill, Felix [2 ]
Rudolph, Maja [3 ]
Baldridge, Jason [4 ]
Schutze, Hinrich [5 ]
机构
[1] Stanford Univ, Dept Psychol, Stanford, CA 94305 USA
[2] DeepMind, London N1C 4AG, England
[3] Bosch Ctr Artificial Intelligence, D-71272 Renningen, Germany
[4] Google Res, Austin, TX 78701 USA
[5] Ludwig Maximilian Univ Munich, Ctr Informat & Language Proc, D-80538 Munich, Germany
基金
欧洲研究理事会;
关键词
natural language understanding; deep learning; situation models; cognitive neuroscience; artificial intelligence; COMPLEMENTARY LEARNING-SYSTEMS; INTERACTIVE ACTIVATION; SEMANTIC KNOWLEDGE; MEMORY; PERCEPTION; REPRESENTATION; PRINCIPLES;
D O I
10.1073/pnas.1910416117
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Language is crucial for human intelligence, but what exactly is its role? We take language to be a part of a system for understanding and communicating about situations. In humans, these abilities emerge gradually from experience and depend on domain-general principles of biological neural networks: connection-based learning, distributed representation, and context-sensitive, mutual constraint satisfaction-based processing. Current artificial language processing systems rely on the same domain general principles, embodied in artificial neural networks. Indeed, recent progress in this field depends on query-based attention, which extends the ability of these systems to exploit context and has contributed to remarkable breakthroughs. Nevertheless, most current models focus exclusively on language-internal tasks, limiting their ability to perform tasks that depend on understanding situations. These systems also lack memory for the contents of prior situations outside of a fixed contextual span. We describe the organization of the brain's distributed understanding system, which includes a fast learning system that addresses the memory problem. We sketch a framework for future models of understanding drawing equally on cognitive neuroscience and artificial intelligence and exploiting query-based attention. We highlight relevant current directions and consider further developments needed to fully capture human-level language understanding in a computational system.
引用
收藏
页码:25966 / 25974
页数:9
相关论文
共 80 条
[1]   The real-time mediation of visual attention by language and world knowledge: Linking anticipatory (and other) eye movements to linguistic processing [J].
Altmann, Gerry T. M. ;
Kamide, Yuki .
JOURNAL OF MEMORY AND LANGUAGE, 2007, 57 (04) :502-518
[2]  
[Anonymous], 2013, NIPS
[3]  
[Anonymous], 1961, PRINCIPLES NEURODYNA, DOI DOI 10.1038/S41586-020-2258-0
[4]  
[Anonymous], 1977, Attention and performance IV pp
[5]  
[Anonymous], 1988, Parallel distributed processing1
[6]  
[Anonymous], 2016, P C ASS MACH TRANSL
[7]  
Bahdanau D., 2016, 3 INT C LEARN REPR
[8]   Discovering Event Structure in Continuous Narrative Perception and Memory [J].
Baldassano, Christopher ;
Chen, Janice ;
Zadbood, Asieh ;
Pillow, Jonathan W. ;
Hasson, Uri ;
Norman, Kenneth A. .
NEURON, 2017, 95 (03) :709-+
[9]   Is relational reasoning dependent on language? A voxel-based lesion symptom mapping study [J].
Baldo, Juliana V. ;
Bunge, Silvia A. ;
Wilson, Stephen M. ;
Dronkers, Nina F. .
BRAIN AND LANGUAGE, 2010, 113 (02) :59-64
[10]  
Bisk Y., 2020, ARXIV200410151