A Quantized Training Framework for Robust and Accurate ReRAM-based Neural Network Accelerators

被引:7
作者
Zhang, Chenguang [1 ]
Zhou, Pingqiang [1 ]
机构
[1] ShanghaiTech Univ, Sch Informat Sci & Technol, Shanghai, Peoples R China
来源
2021 26TH ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC) | 2021年
关键词
ReRAM; Neural Network; Variation; Robust; Quantize;
D O I
10.1145/3394885.3431528
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Neural networks (NN), especially deep neural networks (DNN), have achieved great success in lots of fields. ReRAM crossbar, as a promising candidate, is widely employed to accelerate neural network owing to its nature of processing MVM. However, ReRAM crossbar suffers high conductance variation due to many non-ideal effects, resulting in great inference accuracy degradation. Recent works use uniform quantization to enhance the tolerance of conductance variation, but these methods still suffer high accuracy loss with large variation. In this paper, firstly, we analyze the impact of the quantization and conductance variation on the accuracy. Then, based on two observation, we propose a quantized training framework to enhance the robustness and accuracy of the neural network running on the accelerator, by introducing a smart non-uniform quantizer. This framework consists of a robust trainable quantizer and a corresponding training method, and needs no extra hardware overhead and compatible with a standard neural network training procedure. Experimental results show that our proposed method can improve inference accuracy by 10% similar to 30% under large variation, compared with uniform quantization method.
引用
收藏
页码:43 / 48
页数:6
相关论文
共 50 条
  • [21] 3D-ReG: A 3D ReRAM-based Heterogeneous Architecture for Training Deep Neural Networks
    Li, Bing
    Doppa, Janardhan Rao
    Pande, Partha Pratim
    Chakrabarty, Krishnendu
    Qiu, Joe X.
    Li, Hai
    ACM JOURNAL ON EMERGING TECHNOLOGIES IN COMPUTING SYSTEMS, 2020, 16 (02)
  • [22] Exploiting device-level non-idealities for adversarial attacks on ReRAM-based neural networks
    McLemore, Tyler
    Sunbury, Robert
    Brodzik, Seth
    Cronin, Zachary
    Timmons, Elias
    Chakraborty, Dwaipayan
    Memories - Materials, Devices, Circuits and Systems, 2023, 4
  • [23] A ReRAM-Based Row-Column-Oriented Memory Architecture for Convolutional Neural Networks
    Chen, Yan
    Zhang, Jing
    Xu, Yuebing
    Zhang, Yingjie
    Zhang, Renyuan
    Nakashima, Yasuhiko
    IEICE TRANSACTIONS ON ELECTRONICS, 2019, E102C (07) : 580 - 584
  • [24] A Low Store Energy and Robust ReRAM-Based Flip-Flop for Normally Off Microprocessors
    Chien, Tsai-Kan
    Chiou, Lih-Yih
    Chuang, Yao-Chun
    Sheu, Shyh-Shyuan
    Li, Heng-Yuan
    Wang, Pei-Hua
    Ku, Tzu-Kun
    Tsai, Ming-Jinn
    Wu, Chih-I
    2016 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2016, : 2803 - 2806
  • [25] Energy-Efficient ReRAM-based ML Training via Mixed Pruning and Reconfigurable ADC
    Ogbogu, Chukwufumnanya
    Soumen, Mohapatra
    Joardar, Biresh Kumar
    Doppa, Janardhan Rao
    Heo, Deuk
    Chakrabarty, Krishnendu
    Pande, Partha Pratim
    2023 IEEE/ACM INTERNATIONAL SYMPOSIUM ON LOW POWER ELECTRONICS AND DESIGN, ISLPED, 2023,
  • [26] ReRAM-based graph attention network with node-centric edge searching and hamming similarity
    Mao, Ruibin
    Sheng, Xia
    Graves, Catherine
    Xu, Cong
    Li, Can
    2023 60TH ACM/IEEE DESIGN AUTOMATION CONFERENCE, DAC, 2023,
  • [27] HitM: High-Throughput ReRAM-based PIM for Multi-Modal Neural Networks
    Li, Bing
    Wang, Ying
    Chen, Yiran
    2020 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER AIDED-DESIGN (ICCAD), 2020,
  • [28] Reliability Issues in Analog ReRAM Based Neural-Network Processor
    Yasuhara, Ryutaro
    Ono, Takashi
    Mochida, Reiji
    Muraoka, Shunsaku
    Kouno, Kazuyuki
    Katayama, Koji
    Hayata, Yuriko
    Nakayama, Masayoshi
    Suwa, Hitoshi
    Hayakawa, Yukio
    Mikawa, Takumi
    Gohou, Yasushi
    Yoneda, Shinichi
    2019 IEEE INTERNATIONAL RELIABILITY PHYSICS SYMPOSIUM (IRPS), 2019,
  • [29] ADC-Free ReRAM-Based In-Situ Accelerator for Energy-Efficient Binary Neural Networks
    Kim, Hyeonuk
    Jung, Youngbeom
    Kim, Lee-Sup
    IEEE TRANSACTIONS ON COMPUTERS, 2024, 73 (02) : 353 - 365
  • [30] ReRAM Crossbar based Recurrent Neural Network for Human Activity Detection
    Long, Yun
    Jung, Eui Min
    Kung, Jaeha
    Mukhopadhyay, Saibal
    2016 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2016, : 939 - 946