DARL: Distributed Reconfigurable Accelerator for Hyperdimensional Reinforcement Learning

被引:13
作者
Chen, Hanning [1 ]
Issa, Mariam [1 ]
Ni, Yang [1 ]
Imani, Mohsen [1 ]
机构
[1] Univ Calif Irvine, Irvine, CA 92717 USA
来源
2022 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER AIDED DESIGN, ICCAD | 2022年
基金
美国国家科学基金会;
关键词
D O I
10.1145/3508352.3549437
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Reinforcement Learning (RL) is a powerful technology to solve decisionmaking problems such as robotics control. Modern RL algorithms, i.e., Deep Q-Learning, are based on costly and resource hungry deep neural networks. This motivates us to deploy alternative models for powering RL agents on edge devices. Recently, brain-inspired HyperDimensional Computing (HDC) has been introduced as a promising solution for lightweight and efficient machine learning, particularly for classification. In this work, we develop a novel platform capable of real-time hyper dimensional reinforcement learning. Our heterogeneous CPU-FPGA platform, called DARL, maximizes FPGA's computing capabilities by applying hardware optimizations to hyperdimensional computing's critical operations, including hardware -friendly encoder IP, the hypervector chunk fragmentation, and the delayed model update. Aside from hardware innovation, we also extend the platform to basic single agent RL to support multi-agents distributed learning. We evaluate the effectiveness of our approach on OpenAl Gym tasks. Our results show that the FPGA platform provides on average 20x speedup compared to current state-of-the-art hyperdimensional RL methods running on Intel Xeon 6226 CPU. In addition, DARL provides around 4.8x faster and 4.2x higher energy efficiency compared to the state-of-the-art RL accelerator While ensuring a better or comparable quality of learning.
引用
收藏
页数:9
相关论文
共 47 条
  • [1] [Anonymous], OP GYM LUN
  • [2] [Anonymous], 2017, 2017 IEEE INT C REB, DOI DOI 10.1145/3061639.3062210
  • [3] [Anonymous], OP GYM CARTP V1
  • [4] Reinforcement Learning, Fast and Slow
    Botvinick, Matthew
    Ritter, Sam
    Wang, Jane X.
    Kurth-Nelson, Zeb
    Blundell, Charles
    Hassabis, Demis
    [J]. TRENDS IN COGNITIVE SCIENCES, 2019, 23 (05) : 408 - 422
  • [5] Brockman G., 2016, ARXIV160001540
  • [6] FA3C: FPGA-Accelerated Deep Reinforcement Learning
    Cho, Hyungmin
    Oh, Pyeongseok
    Park, Jiyoung
    Jung, Wookeun
    Lee, Jaejin
    [J]. TWENTY-FOURTH INTERNATIONAL CONFERENCE ON ARCHITECTURAL SUPPORT FOR PROGRAMMING LANGUAGES AND OPERATING SYSTEMS (ASPLOS XXIV), 2019, : 499 - 513
  • [7] Parallel Implementation of Reinforcement Learning Q-Learning Technique for FPGA
    Da Silva, Lucileide M. D.
    Torquato, Matheus F.
    Fernandes, Marcelo A. C.
    [J]. IEEE ACCESS, 2019, 7 : 2782 - 2798
  • [8] Espeholt L, 2018, PR MACH LEARN RES, V80
  • [9] An Introduction to Deep Reinforcement Learning
    Francois-Lavet, Vincent
    Henderson, Peter
    Islam, Riashat
    Bellemare, Marc G.
    Pineau, Joelle
    [J]. FOUNDATIONS AND TRENDS IN MACHINE LEARNING, 2018, 11 (3-4): : 219 - 354
  • [10] Classification Using Hyperdimensional Computing: A Review
    Ge, Lulu
    Parhi, Keshab K.
    [J]. IEEE CIRCUITS AND SYSTEMS MAGAZINE, 2020, 20 (02) : 30 - 47