NPCL: Neural Processes for Uncertainty-Aware Continual Learning

被引:0
作者
Jha, Saurav [1 ]
Gong, Dong [1 ]
Zhao, He [2 ]
Yao, Lina [1 ,2 ]
机构
[1] UNSW Sydney, Sydney, NSW, Australia
[2] CSIROs Data61, Eveleigh, Australia
来源
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023) | 2023年
基金
澳大利亚研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Continual learning (CL) aims to train deep neural networks efficiently on streaming data while limiting the forgetting caused by new tasks. However, learning transferable knowledge with less interference between tasks is difficult, and real-world deployment of CL models is limited by their inability to measure predictive uncertainties. To address these issues, we propose handling CL tasks with neural processes (NPs), a class of meta-learners that encode different tasks into probabilistic distributions over functions all while providing reliable uncertainty estimates. Specifically, we propose an NP-based CL approach (NPCL) with task-specific modules arranged in a hierarchical latent variable model. We tailor regularizers on the learned latent distributions to alleviate forgetting. The uncertainty estimation capabilities of the NPCL can also be used to handle the task head/module inference challenge in CL. Our experiments show that the NPCL outperforms previous CL approaches. We validate the effectiveness of uncertainty estimation in the NPCL for identifying novel data and evaluating instance-level model confidence. Code is available at https://github.com/srvCodes/NPCL.
引用
收藏
页数:25
相关论文
共 55 条
[1]  
[Anonymous], 2018, PMLR
[2]  
[Anonymous], 2018, PMLR
[3]  
[Anonymous], 2017, PMLR
[4]  
Ba J.L., 2016, Layer Normalization
[5]  
Benjamin A., 2018, INT C LEARN REPR
[6]  
Boschini M., 2022, IEEE Transactions on Pattern Analysis and Machine Intelligence
[7]  
Buzzega Pietro, 2020, ADV NEURAL INFORM PR, V33, P15920
[8]  
Chaudhry A., 2020, P ADV NEUR INF PROC, P9900
[9]  
Chaudhry A., 2019, INT C LEARNING REPRE, P1
[10]  
Chaudhry A., 2019, ARXIV