Continual Learning, Fast and Slow

被引:6
作者
Pham, Quang [1 ]
Liu, Chenghao [2 ]
Hoi, Steven C. H. [2 ,3 ]
机构
[1] ASTAR, Inst Infocomm Res I2R, Singapore 138632, Singapore
[2] Salesforce Res Asia, Singapore 038985, Singapore
[3] Singapore Management Univ, Singapore 188065, Singapore
关键词
Continual learning; fast and slow learning; SYSTEMS;
D O I
10.1109/TPAMI.2023.3324203
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
According to the Complementary Learning Systems (CLS) theory (McClelland et al. 1995) in neuroscience, humans do effective continual learning through two complementary systems: a fast learning system centered on the hippocampus for rapid learning of the specifics, individual experiences; and a slow learning system located in the neocortex for the gradual acquisition of structured knowledge about the environment. Motivated by this theory, we propose DualNets (for Dual Networks), a general continual learning framework comprising a fast learning system for supervised learning of pattern-separated representation from specific tasks and a slow learning system for representation learning of task-agnostic general representation via Self-Supervised Learning (SSL). DualNets can seamlessly incorporate both representation types into a holistic framework to facilitate better continual learning in deep neural networks. Via extensive experiments, we demonstrate the promising results of DualNets on a wide range of continual learning protocols, ranging from the standard offline, task-aware setting to the challenging online, task-free scenario. Notably, on the CTrL (Veniat et al. 2020) benchmark that has unrelated tasks with vastly different visual images, DualNets can achieve competitive performance with existing state-of-the-art dynamic architecture strategies (Ostapenko et al. 2021). Furthermore, we conduct comprehensive ablation studies to validate DualNets efficacy, robustness, and scalability.
引用
收藏
页码:134 / 149
页数:16
相关论文
共 91 条
[1]  
Rusu AA, 2016, Arxiv, DOI arXiv:1606.04671
[2]  
Aljundi R., 2019, IEEE C COMP VIS PATT, P11254
[3]  
Aljundi R, 2019, ADV NEUR IN, V32
[4]   Memory Aware Synapses: Learning What (not) to Forget [J].
Aljundi, Rahaf ;
Babiloni, Francesca ;
Elhoseiny, Mohamed ;
Rohrbach, Marcus ;
Tuytelaars, Tinne .
COMPUTER VISION - ECCV 2018, PT III, 2018, 11207 :144-161
[5]  
Arani F., 2022, INT C LEARN REPR, P1
[6]   Remote Memory and the Hippocampus: A Constructive Critique [J].
Barry, Daniel N. ;
Maguire, Eleanor A. .
TRENDS IN COGNITIVE SCIENCES, 2019, 23 (02) :128-142
[7]   Representation Learning: A Review and New Perspectives [J].
Bengio, Yoshua ;
Courville, Aaron ;
Vincent, Pascal .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2013, 35 (08) :1798-1828
[8]  
Bhat PS, 2022, P MACHINE LEARNING R, P390
[9]  
Buzzega P., 2020, PROC INT C NEURAL IN
[10]  
Caccia M, 2021, Arxiv, DOI arXiv:2003.05856