DARL: Distributed Reconfigurable Accelerator for Hyperdimensional Reinforcement Learning

被引:16
作者
Chen, Hanning [1 ]
Issa, Mariam [1 ]
Ni, Yang [1 ]
Imani, Mohsen [1 ]
机构
[1] Univ Calif Irvine, Irvine, CA 92717 USA
来源
2022 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER AIDED DESIGN, ICCAD | 2022年
基金
美国国家科学基金会;
关键词
D O I
10.1145/3508352.3549437
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Reinforcement Learning (RL) is a powerful technology to solve decisionmaking problems such as robotics control. Modern RL algorithms, i.e., Deep Q-Learning, are based on costly and resource hungry deep neural networks. This motivates us to deploy alternative models for powering RL agents on edge devices. Recently, brain-inspired HyperDimensional Computing (HDC) has been introduced as a promising solution for lightweight and efficient machine learning, particularly for classification. In this work, we develop a novel platform capable of real-time hyper dimensional reinforcement learning. Our heterogeneous CPU-FPGA platform, called DARL, maximizes FPGA's computing capabilities by applying hardware optimizations to hyperdimensional computing's critical operations, including hardware -friendly encoder IP, the hypervector chunk fragmentation, and the delayed model update. Aside from hardware innovation, we also extend the platform to basic single agent RL to support multi-agents distributed learning. We evaluate the effectiveness of our approach on OpenAl Gym tasks. Our results show that the FPGA platform provides on average 20x speedup compared to current state-of-the-art hyperdimensional RL methods running on Intel Xeon 6226 CPU. In addition, DARL provides around 4.8x faster and 4.2x higher energy efficiency compared to the state-of-the-art RL accelerator While ensuring a better or comparable quality of learning.
引用
收藏
页数:9
相关论文
共 45 条
[1]  
[Anonymous], 2021, P INT C HIGH PERF CO
[2]  
[Anonymous], 2017, 2017 IEEE INT C REB, DOI DOI 10.1145/3061639.3062210
[3]  
[Anonymous], 2021, P 39 INT C COMP AID, DOI DOI 10.1109/TSMC.2020.2964808
[4]   Reinforcement Learning, Fast and Slow [J].
Botvinick, Matthew ;
Ritter, Sam ;
Wang, Jane X. ;
Kurth-Nelson, Zeb ;
Blundell, Charles ;
Hassabis, Demis .
TRENDS IN COGNITIVE SCIENCES, 2019, 23 (05) :408-422
[5]  
Brockman G., 2016, ARXIV160001540
[6]   FA3C: FPGA-Accelerated Deep Reinforcement Learning [J].
Cho, Hyungmin ;
Oh, Pyeongseok ;
Park, Jiyoung ;
Jung, Wookeun ;
Lee, Jaejin .
TWENTY-FOURTH INTERNATIONAL CONFERENCE ON ARCHITECTURAL SUPPORT FOR PROGRAMMING LANGUAGES AND OPERATING SYSTEMS (ASPLOS XXIV), 2019, :499-513
[7]   Parallel Implementation of Reinforcement Learning Q-Learning Technique for FPGA [J].
Da Silva, Lucileide M. D. ;
Torquato, Matheus F. ;
Fernandes, Marcelo A. C. .
IEEE ACCESS, 2019, 7 :2782-2798
[8]  
Espeholt L, 2018, PR MACH LEARN RES, V80
[9]   An Introduction to Deep Reinforcement Learning [J].
Francois-Lavet, Vincent ;
Henderson, Peter ;
Islam, Riashat ;
Bellemare, Marc G. ;
Pineau, Joelle .
FOUNDATIONS AND TRENDS IN MACHINE LEARNING, 2018, 11 (3-4) :219-354
[10]   Classification Using Hyperdimensional Computing: A Review [J].
Ge, Lulu ;
Parhi, Keshab K. .
IEEE CIRCUITS AND SYSTEMS MAGAZINE, 2020, 20 (02) :30-47