Exploring Model Stability of Deep Neural Networks for Reliable RRAM-Based In-Memory Acceleration

被引:5
作者
Krishnan, Gokul [1 ]
Yang, Li [1 ]
Sun, Jingbo [1 ]
Hazra, Jubin [2 ]
Du, Xiaocong [1 ]
Liehr, Maximilian [2 ]
Li, Zheng [1 ]
Beckmann, Karsten [2 ]
Joshi, Rajiv, V [3 ]
Cady, Nathaniel C. [2 ]
Fan, Deliang [1 ]
Cao, Yu [1 ]
机构
[1] Arizona State Univ, Sch Elect Comp & Energy Engn, Tempe, AZ 85287 USA
[2] State Univ New York Polytech, Albany, NY 12246 USA
[3] IBM Corp, TJ Watson Res Ctr, Yorktown Hts, NY 10598 USA
关键词
Stability analysis; Computational modeling; Quantization (signal); Semiconductor device modeling; Training; Perturbation methods; Neural networks; In-memory computing; RRAM; model stability; deep neural networks; reliability; pruning; quantization;
D O I
10.1109/TC.2022.3174585
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
RRAM-based in-memory computing (IMC) effectively accelerates deep neural networks (DNNs). Furthermore, model compression techniques, such as quantization and pruning, are necessary to improve algorithm mapping and hardware performance. However, in the presence of RRAM device variations, low-precision and sparse DNNs suffer from severe post-mapping accuracy loss. To address this, in this work, we investigate a new metric, model stability, from the loss landscape to help shed light on accuracy loss under variations and model compression, which guides an algorithmic solution to maximize model stability and mitigate accuracy loss. Based on statistical data from a CMOS/RRAM 1T1R test chip at 65nm, we characterize wafer-level RRAM variations and develop a cross-layer benchmark tool that incorporates quantization, pruning, device variations, model stability, and IMC architecture parameters to assess post-mapping accuracy and hardware performance. Leveraging this tool, we show that a loss-landscape-based DNN model selection for stability effectively tolerates device variations and achieves a post-mapping accuracy higher than that with 50% lower RRAM variations. Moreover, we quantitatively interpret why model pruning increases the sensitivity to variations, while a lower-precision model has better tolerance to variations. Finally, we propose a novel variation-aware training method to improve model stability, in which there exists the most stable model for the best post-mapping accuracy of compressed DNNs. Experimental evaluation of the method shows up to 19%, 21%, and 11% post-mapping accuracy improvement for our 65nm RRAM device, across various precision and sparsity, on CIFAR-10, CIFAR-100, and SVHN datasets, respectively.
引用
收藏
页码:2740 / 2752
页数:13
相关论文
共 41 条
  • [1] GENIEx: A Generalized Approach to Emulating Non-Ideality in Memristive Xbars using Neural Networks
    Chakraborty, Indranil
    Ali, Mustafa Fayez
    Kim, Dong Eun
    Ankit, Aayush
    Roy, Kaushik
    [J]. PROCEEDINGS OF THE 2020 57TH ACM/EDAC/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2020,
  • [2] Charan G., 2020, PROC 57 ACMIEEE AUTO, P1
  • [3] RRAM Defect Modeling and Failure Analysis Based on March Test and a Novel Squeeze-Search Scheme
    Chen, Ching-Yi
    Shih, Hsiu-Chuan
    Wu, Cheng-Wen
    Lin, Chih-He
    Chiu, Pi-Feng
    Sheu, Shyh-Shyuan
    Chen, Frederick T.
    [J]. IEEE TRANSACTIONS ON COMPUTERS, 2015, 64 (01) : 180 - 190
  • [4] Chen LR, 2017, DES AUT TEST EUROPE, P19, DOI 10.23919/DATE.2017.7926952
  • [5] Eyeriss v2: A Flexible Accelerator for Emerging Deep Neural Networks on Mobile Devices
    Chen, Yu-Hsin
    Yange, Tien-Ju
    Emer, Joel S.
    Sze, Vivienne
    [J]. IEEE JOURNAL ON EMERGING AND SELECTED TOPICS IN CIRCUITS AND SYSTEMS, 2019, 9 (02) : 292 - 308
  • [6] Noise-based Selection of Robust Inherited Model for Accurate Continual Learning
    Du, Xiaocong
    Li, Zheng
    Seo, Jae-sun
    Liu, Frank
    Cao, Yu
    [J]. 2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2020), 2020, : 983 - 988
  • [7] Frankle J., 2019, PROC INT C LEARN REP
  • [8] Han S, 2015, ADV NEUR IN, V28
  • [9] Deep Residual Learning for Image Recognition
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 770 - 778
  • [10] He Z., 2019, PROC 56 ACMIEEE AUTO, P1