Enabling Continual Learning with Differentiable Hebbian Plasticity

被引:4
作者
Thangarasa, Vithursan [1 ]
Miconi, Thomas [2 ]
Taylor, Graham W. [1 ]
机构
[1] Univ Guelph, Sch Engn, Vector Inst Artificial Intelligence, Guelph, ON, Canada
[2] Uber AI, San Francisco, CA USA
来源
2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN) | 2020年
关键词
neural networks; plasticity; catastrophic forgetting; continual learning. hebbian learning;
D O I
10.1109/ijcnn48605.2020.9206764
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Continual learning is the problem of sequentially learning new tasks or knowledge while protecting previously acquired knowledge. However, catastrophic forgetting poses a grand challenge for neural networks performing such learning process. Thus, neural networks that are deployed in the real world often struggle in scenarios where the data distribution is non-stationary (concept drift), imbalanced, or not always fully available, i.e., rare edge cases. We propose a Differentiable Hebbian Consolidation model which is composed of a Differentiable Hebbian Plasticity (DHP) Softmax layer that adds a rapid learning plastic component (compressed episodic memory) to the fixed (slow changing) parameters of the softmax output layer; enabling learned representations to be retained for a longer timescale. We demonstrate the flexibility of our method by integrating well-known task-specific synaptic consolidation methods to penalize changes in the slow weights that are important for each target task. We evaluate our approach on the Permuted MNIST, Split MNIST and Vision Datasets Mixture benchmarks, and introduce an imbalanced variant of Permuted MNIST - a dataset that combines the challenges of class imbalance and concept drift. Our proposed model requires no additional hyperparameters and outperforms comparable baselines by reducing forgetting.
引用
收藏
页数:8
相关论文
共 48 条
  • [1] Memory retention - the synaptic stability versus plasticity dilemma
    Abraham, WC
    Robins, A
    [J]. TRENDS IN NEUROSCIENCES, 2005, 28 (02) : 73 - 78
  • [2] Memory Aware Synapses: Learning What (not) to Forget
    Aljundi, Rahaf
    Babiloni, Francesca
    Elhoseiny, Mohamed
    Rohrbach, Marcus
    Tuytelaars, Tinne
    [J]. COMPUTER VISION - ECCV 2018, PT III, 2018, 11207 : 144 - 161
  • [3] [Anonymous], 2013, P 30 INT C MACH LEAR
  • [4] [Anonymous], COMPUTER VISION GRAP
  • [5] [Anonymous], 1998, Lifelong learning algorithms
  • [6] Ans B., 2004, CONNECTION SCI
  • [7] Atkinson C., 2018, CORR
  • [8] Ba J. L., 2016, LAYER NORMALIZATION, Vabs/1607.06450
  • [9] Bailey C. H., 2015, COLD SPRING HARBOR P
  • [10] Benna M. K., 2016, NATURE NEUROSCIENCE