Saving Energy of RRAM-Based Neural Accelerator Through State-Aware Computing

被引:1
|
作者
He, Yintao [1 ,2 ]
Wang, Ying [1 ,2 ]
Li, Huawei [1 ,2 ,3 ]
Li, Xiaowei [1 ,2 ]
机构
[1] Chinese Acad Sci, Inst Comp Technol, State Key Lab Comp Architecture, Beijing 100190, Peoples R China
[2] Univ Chinese Acad Sci, Beijing 100190, Peoples R China
[3] Peng Cheng Lab, Shenzhen 518066, Peoples R China
基金
中国国家自然科学基金;
关键词
Computer architecture; Microprocessors; Resistance; Power demand; Training; Biological neural networks; Optimization; Low power (LP); neural networks; processing-in-memory; resistive random-access memory (RRAM);
D O I
10.1109/TCAD.2021.3103147
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
In-memory computing (IMC) is recognized as one of the most promising architecture solution to realize energy-efficient neural network inference. Amongst many memory technology, resistive RAM (RRAM) is a very attractive device to implement the IMC-based neural network accelerator architecture, which is particularly suitable for power-constrained IoT systems. Due to the nature of low leakage and in-situ computing, the dynamic power consumption of dot-production operations in RRAM crossbars dominates the chip power, especially when applied to low-precision neural networks. This work investigates the correlation between the cell resistance state and the crossbar operation power, and proposes a state-aware RRAM accelerator (SARA) architecture for energy-efficient low-precision neural networks. With the proposed state-aware network training and mapping strategy, crossbars in the RRAM accelerator can perform in a lower power state. Furthermore, we also leverage the proposed RRAM accelerator architecture to reduce the power consumption of high-precision network inference with both single-level or multilevel RRAM. The evaluation results show that for binary neural networks, our design saves 40.53% RRAM computing energy on average over the baseline. For high precision neural networks, the proposed method reduces 11.67% computing energy on average without any accuracy loss.
引用
收藏
页码:2115 / 2127
页数:13
相关论文
共 50 条
  • [1] R-Accelerator: An RRAM-Based CGRA Accelerator With Logic Contraction
    Chen, Zhengyu
    Zhou, Hai
    Gu, Jie
    IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS, 2019, 27 (11) : 2655 - 2667
  • [2] RRAM-Based Analog Approximate Computing
    Li, Boxun
    Gu, Peng
    Shan, Yi
    Wang, Yu
    Chen, Yiran
    Yang, Huazhong
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2015, 34 (12) : 1905 - 1917
  • [3] BitS-Net: Bit-Sparse Deep Neural Network for Energy-Efficient RRAM-Based Compute-In-Memory
    Karimzadeh, Foroozan
    Yoon, Jong-Hyeok
    Raychowdhury, Arijit
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2022, 69 (05) : 1952 - 1961
  • [4] Nonlinear Weight Quantification for Mitigating Stress Induced Disturb Effect on Multilevel RRAM-Based Neural Network Accelerator
    Wu, Lindong
    Yu, Zhizhen
    Qin, Yabo
    Chen, Qingyu
    Cai, Yimao
    Huang, Ru
    IEEE JOURNAL OF THE ELECTRON DEVICES SOCIETY, 2021, 9 : 1257 - 1261
  • [5] An RRAM-Based Digital Computing-in-Memory Macro With Dynamic Voltage Sense Amplifier and Sparse-Aware Approximate Adder Tree
    He, Yifan
    Yue, Jinshan
    Feng, Xiaoyu
    Huang, Yuxuan
    Jia, Hongyang
    Wang, Jingyu
    Zhang, Lu
    Sun, Wenyu
    Yang, Huazhong
    Liu, Yongpan
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2023, 70 (02) : 416 - 420
  • [6] Logic Synthesis for RRAM-Based In-Memory Computing
    Shirinzadeh, Saeideh
    Soeken, Mathias
    Gaillardon, Pierre-Emmanuel
    Drechsler, Rolf
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2018, 37 (07) : 1422 - 1435
  • [7] Variation Enhanced Attacks Against RRAM-Based Neuromorphic Computing System
    Lv, Hao
    Li, Bing
    Zhang, Lei
    Liu, Cheng
    Wang, Ying
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2023, 42 (05) : 1588 - 1596
  • [8] A Hardware Friendly Variation-Tolerant Framework for RRAM-Based Neuromorphic Computing
    Gu, Fang-Yi
    Yang, Cheng-Han
    Lin, Ing-Chao
    Chang, Da-Wei
    Lu, Darsen D.
    Schlichtmann, Ulf
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2024, 71 (12) : 6419 - 6432
  • [9] Rescuing RRAM-Based Computing From Static and Dynamic Faults
    Lin, Jilan
    Wen, Cheng-Da
    Hu, Xing
    Tang, Tianqi
    Lin, Ing-Chao
    Wang, Yu
    Xie, Yuan
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2021, 40 (10) : 2049 - 2062
  • [10] Accurate and Energy-Efficient Bit-Slicing for RRAM-Based Neural Networks
    Diware, Sumit
    Singh, Abhairaj
    Gebregiorgis, Anteneh
    Joshi, Rajiv V.
    Hamdioui, Said
    Bishnoi, Rajendra
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2023, 7 (01): : 164 - 177