Comparing continual task learning in minds and machines

被引:79
作者
Flesch, Timo [1 ]
Balaguer, Jan [1 ,2 ]
Dekker, Ronald [1 ]
Nili, Hamed [1 ]
Summerfield, Christopher [1 ,2 ]
机构
[1] Univ Oxford, Dept Expt Psychol, Oxford OX2 6BW, England
[2] DeepMind, London EC4A 3TW, England
基金
欧洲研究理事会;
关键词
continual learning; catastrophic forgetting; categorization; task factorization; representational similarity analysis; BAYESIAN MODEL SELECTION; CONNECTIONIST MODELS; CATEGORIZATION; SYSTEMS; MEMORY; ORDER; CONSTRAINTS; COGNITION; NETWORKS; SEE;
D O I
10.1073/pnas.1800755115
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Humans can learn to perform multiple tasks in succession over the lifespan ("continual" learning), whereas current machine learning systems fail. Here, we investigated the cognitive mechanisms that permit successful continual learning in humans and harnessed our behavioral findings for neural network design. Humans categorized naturalistic images of trees according to one of two orthogonal task rules that were learned by trial and error. Training regimes that focused on individual rules for prolonged periods (blocked training) improved human performance on a later test involving randomly interleaved rules, compared with control regimes that trained in an interleaved fashion. Analysis of human error patterns suggested that blocked training encouraged humans to form "factorized" representation that optimally segregated the tasks, especially for those individuals with a strong prior bias to represent the stimulus space in a well-structured way. By contrast, standard supervised deep neural networks trained on the same tasks suffered catastrophic forgetting under blocked training, due to representational interference in the deeper layers. However, augmenting deep networks with an unsupervised generative model that allowed it to first learn a good embedding of the stimulus space (similar to that observed in humans) reduced catastrophic forgetting under blocked training. Building artificial agents that first learn a model of the world may be one promising route to solving continual task performance in artificial intelligence research.
引用
收藏
页码:E10313 / E10322
页数:10
相关论文
共 48 条
[1]  
[Anonymous], 2015, PRIORITIZED EXPERIEN
[2]   Human category learning [J].
Ashby, EG ;
Maddox, WT .
ANNUAL REVIEW OF PSYCHOLOGY, 2005, 56 :149-178
[3]   Perception of Object Persistence: The Origins of Object Permanence in Infancy [J].
Bremner, J. Gavin ;
Slater, Alan M. ;
Johnson, Scott P. .
CHILD DEVELOPMENT PERSPECTIVES, 2015, 9 (01) :7-13
[4]   What you learn is more than what you see: what can sequencing effects tell us about inductive category learning? [J].
Carvalho, Paulo F. ;
Goldstone, Robert L. .
FRONTIERS IN PSYCHOLOGY, 2015, 6
[5]   Putting category learning in order: Category structure and temporal arrangement affect the benefit of interleaved over blocked study [J].
Carvalho, Paulo F. ;
Goldstone, Robert L. .
MEMORY & COGNITION, 2014, 42 (03) :481-495
[6]   The Cost of Structure Learning [J].
Collins, Anne G. E. .
JOURNAL OF COGNITIVE NEUROSCIENCE, 2017, 29 (10) :1646-1655
[7]   Cognitive Control Over Learning: Creating, Clustering, and Generalizing Task-Set Structure [J].
Collins, Anne G. E. ;
Frank, Michael J. .
PSYCHOLOGICAL REVIEW, 2013, 120 (01) :190-229
[8]   VBA: A Probabilistic Treatment of Nonlinear Models for Neurobiological and Behavioural Data [J].
Daunizeau, Jean ;
Adam, Vincent ;
Rigoux, Lionel .
PLOS COMPUTATIONAL BIOLOGY, 2014, 10 (01)
[9]   Catastrophic forgetting in connectionist networks [J].
French, RM .
TRENDS IN COGNITIVE SCIENCES, 1999, 3 (04) :128-135
[10]   Why neurons mix: high dimensionality for higher cognition [J].
Fusi, Stefano ;
Miller, Earl K. ;
Rigotti, Mattia .
CURRENT OPINION IN NEUROBIOLOGY, 2016, 37 :66-74