Enabling Continual Learning with Differentiable Hebbian Plasticity

被引:4
作者
Thangarasa, Vithursan [1 ]
Miconi, Thomas [2 ]
Taylor, Graham W. [1 ]
机构
[1] Univ Guelph, Sch Engn, Vector Inst Artificial Intelligence, Guelph, ON, Canada
[2] Uber AI, San Francisco, CA USA
来源
2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN) | 2020年
关键词
neural networks; plasticity; catastrophic forgetting; continual learning. hebbian learning;
D O I
10.1109/ijcnn48605.2020.9206764
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Continual learning is the problem of sequentially learning new tasks or knowledge while protecting previously acquired knowledge. However, catastrophic forgetting poses a grand challenge for neural networks performing such learning process. Thus, neural networks that are deployed in the real world often struggle in scenarios where the data distribution is non-stationary (concept drift), imbalanced, or not always fully available, i.e., rare edge cases. We propose a Differentiable Hebbian Consolidation model which is composed of a Differentiable Hebbian Plasticity (DHP) Softmax layer that adds a rapid learning plastic component (compressed episodic memory) to the fixed (slow changing) parameters of the softmax output layer; enabling learned representations to be retained for a longer timescale. We demonstrate the flexibility of our method by integrating well-known task-specific synaptic consolidation methods to penalize changes in the slow weights that are important for each target task. We evaluate our approach on the Permuted MNIST, Split MNIST and Vision Datasets Mixture benchmarks, and introduce an imbalanced variant of Permuted MNIST - a dataset that combines the challenges of class imbalance and concept drift. Our proposed model requires no additional hyperparameters and outperforms comparable baselines by reducing forgetting.
引用
收藏
页数:8
相关论文
共 48 条
  • [31] Continual lifelong learning with neural networks: A review
    Parisi, German I.
    Kemker, Ronald
    Part, Jose L.
    Kanan, Christopher
    Wermter, Stefan
    [J]. NEURAL NETWORKS, 2019, 113 : 54 - 71
  • [32] Paulsen O., 2000, CURRENT OPINION NEUR
  • [33] Rae JW, 2018, PR MACH LEARN RES, V80
  • [34] Ring Mark B., 1994, THESIS
  • [35] Ritter H, 2018, Advances in Neural Information Processing Systems
  • [36] Robins A., 1995, Connection Science, V7, P123, DOI 10.1080/09540099550039318
  • [37] Complementary learning systems within the hippocampus: a neural network modelling approach to reconciling episodic memory with statistical learning
    Schapiro, Anna C.
    Turk-Browne, Nicholas B.
    Botvinick, Matthew M.
    Norman, Kenneth A.
    [J]. PHILOSOPHICAL TRANSACTIONS OF THE ROYAL SOCIETY B-BIOLOGICAL SCIENCES, 2017, 372 (1711)
  • [38] Schwarz J, 2018, PR MACH LEARN RES, V80
  • [39] Shin H, 2017, ADV NEUR IN, V30
  • [40] Song S., 2000, NATURE NEUROSCIENCE