MINT: Mixed-precision RRAM-based IN-memory Training Architecture

被引:27
|
作者
Jiang, Hongwu [1 ]
Huang, Shanshi [1 ]
Peng, Xiaochen [1 ]
Yu, Shimeng [1 ]
机构
[1] Georgia Inst Technol, Sch Elect & Comp Engn, Atlanta, GA 30332 USA
关键词
RRAM; compute-in-memory; deep neural network; hardware accelerator;
D O I
10.1109/iscas45731.2020.9181020
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
On-chip training of large-scale deep neural networks (DNNs) is challenging. To solve the memory wall problem, compute-in-memory (CIM) is a promising approach that exploits the analog computation inside the memory array to speed up the vector-matrix multiplication (VMM). Challenges for on-chip CIM training include higher weight precision and higher analog-to-digital converter (ADC) resolution. In this work, we propose a mixed-precision RRAM-based CIM architecture that overcomes these challenges and supports on-chip training. In particular, we split the multi-bit weight into the most significant bits (MSBs) and the least significant bits (LSBs). The forward and backward propagations are performed with CIM transposable arrays for MSBs only, while the weight update is performed in regular memory arrays that store LSBs. Impact of ADC resolution on training accuracy is analyzed. We explore the training performance of a convolutional VGG-like network on the CIFAR-10 dataset using this Mixed-precision IN-memory Training architecture, namely MINT, showing that it can achieve similar to 91% accuracy under hardware constraints and similar to 4.46TOPS/W energy efficiency. Compared with the baseline CIM architectures based on RRAM, it can achieve 1.35x higher energy efficiency and only 31.9% chip size (similar to 98.86 mm(2) at 32 nm node).
引用
收藏
页数:5
相关论文
共 50 条
  • [21] Energy-Efficient SNN Implementation Using RRAM-Based Computation In-Memory (CIM)
    El Arrassi, Asmae
    Gebregiorgis, Anteneh
    El Haddadi, Anass
    Hamdioui, Said
    PROCEEDINGS OF THE 2022 IFIP/IEEE 30TH INTERNATIONAL CONFERENCE ON VERY LARGE SCALE INTEGRATION (VLSI-SOC), 2022,
  • [22] A RRAM-based Associative Memory Cell
    Pan, Yihan
    Foster, Patrick
    Serb, Alex
    Prodromakis, Themis
    2021 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2021,
  • [23] Exploring Model Stability of Deep Neural Networks for Reliable RRAM-Based In-Memory Acceleration
    Krishnan, Gokul
    Yang, Li
    Sun, Jingbo
    Hazra, Jubin
    Du, Xiaocong
    Liehr, Maximilian
    Li, Zheng
    Beckmann, Karsten
    Joshi, Rajiv, V
    Cady, Nathaniel C.
    Fan, Deliang
    Cao, Yu
    IEEE TRANSACTIONS ON COMPUTERS, 2022, 71 (11) : 2740 - 2752
  • [24] Mixed-Precision Deep Learning Based on Computational Memory
    Nandakumar, S. R.
    Le Gallo, Manuel
    Piveteau, Christophe
    Joshi, Vinay
    Mariani, Giovanni
    Boybat, Irem
    Karunaratne, Geethan
    Khaddam-Aljameh, Riduan
    Egger, Urs
    Petropoulos, Anastasios
    Antonakopoulos, Theodore
    Rajendran, Bipin
    Sebastian, Abu
    Eleftheriou, Evangelos
    FRONTIERS IN NEUROSCIENCE, 2020, 14
  • [25] SIMPLY: Design of a RRAM-Based Smart Logic-in-Memory Architecture using RRAM Compact Model
    Puglisi, Francesco Maria
    Zanotti, Tommaso
    Pavan, Paolo
    49TH EUROPEAN SOLID-STATE DEVICE RESEARCH CONFERENCE (ESSDERC 2019), 2019, : 130 - 133
  • [26] Temperature-Dependent Accuracy Analysis and Resistance Temperature Correction in RRAM-Based In-Memory Computing
    Ling, Yaotian
    Wang, Zongwei
    Yu, Zhizhen
    Bao, Shengyu
    Yang, Yuhang
    Bao, Lin
    Sun, Yining
    Cai, Yimao
    Huang, Ru
    IEEE TRANSACTIONS ON ELECTRON DEVICES, 2024, 71 (01) : 294 - 300
  • [27] Fast Logic Synthesis for RRAM-based In-Memory Computing using Majority-Inverter Graphs
    Shirinzadeh, Saeideh
    Soeken, Mathias
    Gaillardon, Pierre-Emmanuel
    Drechsler, Rolf
    PROCEEDINGS OF THE 2016 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE), 2016, : 948 - 953
  • [28] Attar: RRAM-based in-memory attention accelerator with software-hardware co-optimization
    Bing LI
    Ying QI
    Ying WANG
    Yinhe HAN
    Science China(Information Sciences), 2025, 68 (03) : 371 - 387
  • [29] Attar: RRAM-based in-memory attention accelerator with software-hardware co-optimization
    Li, Bing
    Qi, Ying
    Wang, Ying
    Han, Yinhe
    SCIENCE CHINA-INFORMATION SCIENCES, 2025, 68 (03)
  • [30] An isolated symmetrical 2T2R cell enabling high precision and high density for RRAM-based in-memory computing
    Ling, Yaotian
    Wang, Zongwei
    Yang, Yuhang
    Bao, Lin
    Bao, Shengyu
    Wang, Qishen
    Cai, Yimao
    Huang, Ru
    SCIENCE CHINA-INFORMATION SCIENCES, 2024, 67 (05)