Chameleon: Dual Memory Replay for Online Continual Learning on Edge Devices

被引:1
作者
Aggarwal, Shivam [1 ,2 ]
Binici, Kuluhan [2 ,3 ]
Mitra, Tulika [1 ,2 ]
机构
[1] Natl Univ Singapore, Dept Comp Sci, Singapore, Singapore
[2] Natl Univ Singapore, Sch Comp, Singapore, Singapore
[3] ASTAR, Inst Infocomm Res, Singapore, Singapore
基金
新加坡国家研究基金会;
关键词
Continual learning (CL); memory management; model personalization;
D O I
10.1109/TCAD.2023.3347640
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Once deployed on edge devices, a deep neural network model should dynamically adapt to newly discovered environments and personalize its utility for each user. The system must be capable of continual learning (CL), i.e., learning new information from a temporal stream of data in situ without forgetting previously acquired knowledge. However, creating a personalized CL framework poses significant challenges due to limited compute and storage resources on edge devices. Existing methods rely on large memory storage to preserve past data while learning from incoming streams, making them impractical for such devices. In this article, we propose Chameleon as a hardware-friendly CL solution for user-centric CL with dual replay buffers. The strategy takes advantage of the hierarchical memory structure commonly found in edge devices, utilizing a short-term replay store in on-chip memory and a long-term replay store in off-chip memory. We also present an FPGA-based analytical model to estimate the compute and communication costs of the dual replay strategy on the hardware, making effective design choices considering various latent layer options. We conduct extensive experiments on four different models, demonstrating our method's consistent performance across diverse model architectures. Our method achieves up to 7 x speedup and improved energy efficiency on popular edge devices, including ZCU102 FPGA, NVIDIA Jetson Nano, and Google's EdgeTPU. Our code is available at https://github.com/ecolab-nus/Chameleon.
引用
收藏
页码:1663 / 1676
页数:14
相关论文
共 47 条
[1]   Chameleon: Dual Memory Replay for Online Continual Learning on Edge Devices [J].
Aggarwal, Shivam ;
Binici, Kuluhan ;
Mitra, Tulika .
2023 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION, DATE, 2023,
[2]  
Aljundi R, 2019, 33 C NEURAL INFORM P, V32
[3]  
Binici K, 2022, AAAI CONF ARTIF INTE, P6089
[4]  
Borsos Z., 2020, Advances in neural information processing systems, V33, P14879, DOI [10.48550/arXiv.2006.03875, DOI 10.48550/ARXIV.2006.03875]
[5]  
Buzzega P., 2020, Advances in neural information processing systems, V33, P15920
[6]  
Chaudhry A., 2019, arXiv
[7]   Riemannian Walk for Incremental Learning: Understanding Forgetting and Intransigence [J].
Chaudhry, Arslan ;
Dokania, Puneet K. ;
Ajanthan, Thalaiyasingam ;
Torr, Philip H. S. .
COMPUTER VISION - ECCV 2018, PT XI, 2018, 11215 :556-572
[8]  
Chrysakis A, 2020, PR MACH LEARN RES, V119
[9]   uSystolic: Byte-Crawling Unary Systolic Array [J].
Di Wu ;
San Miguel, Joshua .
2022 IEEE INTERNATIONAL SYMPOSIUM ON HIGH-PERFORMANCE COMPUTER ARCHITECTURE (HPCA 2022), 2022, :12-24
[10]  
Farquhar S, 2019, Arxiv, DOI arXiv:1805.09733