Efficient continual learning at the edge with progressive segmented training

被引:1
作者
Du, Xiaocong [1 ]
Venkataramanaiah, Shreyas Kolala [1 ]
Li, Zheng [2 ]
Suh, Han-Sok [1 ]
Yin, Shihui [1 ]
Krishnan, Gokul [1 ]
Liu, Frank [3 ]
Seo, Jae-sun [1 ]
Cao, Yu [1 ,2 ]
机构
[1] Arizona State Univ, Sch Elect, Comp & Energy Engn, Tempe, AZ 85287 USA
[2] Arizona State Univ, Sch Comp Informat & Decis Syst Engn, Tempe, AZ 85287 USA
[3] Oak Ridge Natl Lab, Comp Sci & Math Div, Oak Ridge, TN 37830 USA
来源
NEUROMORPHIC COMPUTING AND ENGINEERING | 2022年 / 2卷 / 04期
关键词
continual learning; acquisitive learning; deep neural network; brain inspiration; model adaption;
D O I
10.1088/2634-4386/ac9899
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
There is an increasing need for continual learning in dynamic systems at the edge, such as self-driving vehicles, surveillance drones, and robotic systems. Such a system requires learning from the data stream, training the model to preserve previous information and adapt to a new task, and generating a single-headed vector for future inference, within a limited power budget. Different from previous continual learning algorithms with dynamic structures, this work focuses on a single network and model segmentation to mitigate catastrophic forgetting problem. Leveraging the redundant capacity of a single network, model parameters for each task are separated into two groups: one important group which is frozen to preserve current knowledge, and a secondary group to be saved (not pruned) for future learning. A fixed-size memory containing a small amount of previously seen data is further adopted to assist the training. Without additional regularization, the simple yet effective approach of progressive segmented training (PST) successfully incorporates multiple tasks and achieves state-of-the-art accuracy in the single-head evaluation on the CIFAR-10 and CIFAR-100 datasets. Moreover, the segmented training significantly improves computation efficiency in continual learning and thus, enabling efficient continual learning at the edge. On Intel Stratix-10 MX FPGA, we further demonstrate the efficiency of PST with representative CNNs trained on CIFAR-10.
引用
收藏
页数:12
相关论文
共 49 条
  • [1] Rusu AA, 2016, Arxiv, DOI arXiv:1606.04671
  • [2] Memory Aware Synapses: Learning What (not) to Forget
    Aljundi, Rahaf
    Babiloni, Francesca
    Elhoseiny, Mohamed
    Rohrbach, Marcus
    Tuytelaars, Tinne
    [J]. COMPUTER VISION - ECCV 2018, PT III, 2018, 11207 : 144 - 161
  • [3] Boski M, 2017, 2017 10TH INTERNATIONAL WORKSHOP ON MULTIDIMENSIONAL (ND) SYSTEMS (NDS)
  • [4] End-to-End Incremental Learning
    Castro, Francisco M.
    Marin-Jimenez, Manuel J.
    Guil, Nicolas
    Schmid, Cordelia
    Alahari, Karteek
    [J]. COMPUTER VISION - ECCV 2018, PT XII, 2018, 11216 : 241 - 257
  • [5] Chaudhry A, 2018, Arxiv, DOI [arXiv:1801.10112, DOI 10.1007/978-3-030-01252-633]
  • [6] Chaudhry A, 2019, Arxiv, DOI arXiv:1812.00420
  • [7] Deo Manish., 2016, Intel White Paper
  • [8] Dhar P, 2019, Arxiv, DOI arXiv:1811.08051
  • [9] Du XC, 2019, Arxiv, DOI arXiv:1905.11533
  • [10] Farquhar S, 2019, Arxiv, DOI arXiv:1805.09733