Tolerating Defects in Low-Power Neural Network Accelerators Via Retraining-Free Weight Approximation

被引:7
作者
Hosseini, Fateme S. [1 ]
Meng, Fanruo [1 ]
Yang, Chengmo [1 ]
Wen, Wujie [2 ]
Cammarota, Rosario [3 ]
机构
[1] Univ Delaware, Dept Elect & Comp Engn, Newark, DE 19716 USA
[2] Lehigh Univ, Dept Elect & Comp Engn, Bethlehem, PA 18015 USA
[3] Intel, San Jose, CA 95134 USA
基金
美国国家科学基金会;
关键词
Neural network accelerator; defect tolerance; reliability; memory faults; approximation; MEMORY; NONVOLATILE; HARD; PCM;
D O I
10.1145/3477016
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Hardware accelerators are essential to the accommodation of ever-increasing Deep Neural Network (DNN) workloads on the resource-constrained embedded devices. While accelerators facilitate fast and energy-efficient DNN operations, their accuracy is threatened by faults in their on-chip and off-chip memories, where millions of DNN weights are held. The use of emerging Non-Volatile Memories (NVM) further exposes DNN accelerators to a non-negligible rate of permanent defects due to immature fabrication, limited endurance, and aging. To tolerate defects in NVM-based DNN accelerators, previous work either requires extra redundancy in hardware or performs defect-aware retraining, imposing significant overhead. In comparison, this paper proposes a set of algorithms that exploit the flexibility in setting the fault-free bits in weight memory to effectively approximate weight values, so as to mitigate defect-induced accuracy drop. These algorithms can be applied as a one-step solution when loading the weights to embedded devices. They only require trivial hardware support and impose negligible run-time overhead. Experiments on popular DNN models show that the proposed techniques successfully boost inference accuracy even in the face of elevated defect rates in the weight memory.
引用
收藏
页数:21
相关论文
共 52 条
[1]  
[Anonymous], 2017, 2017 IEEEACM INT S L, DOI 10.1109/ISLPED.2017.8009163
[2]  
Beckmann K, 2016, MRS ADV, V1, P3355, DOI 10.1557/adv.2016.377
[3]   Experimental Demonstration and Tolerancing of a Large-Scale Neural Network (165 000 Synapses) Using Phase-Change Memory as the Synaptic Weight Element [J].
Burr, Geoffrey W. ;
Shelby, Robert M. ;
Sidler, Severin ;
di Nolfo, Carmelo ;
Jang, Junwoo ;
Boybat, Irem ;
Shenoy, Rohit S. ;
Narayanan, Pritish ;
Virwani, Kumar ;
Giacometti, Emanuele U. ;
Kuerdi, Bulent N. ;
Hwang, Hyunsang .
IEEE TRANSACTIONS ON ELECTRON DEVICES, 2015, 62 (11) :3498-3507
[4]   RRAM Defect Modeling and Failure Analysis Based on March Test and a Novel Squeeze-Search Scheme [J].
Chen, Ching-Yi ;
Shih, Hsiu-Chuan ;
Wu, Cheng-Wen ;
Lin, Chih-He ;
Chiu, Pi-Feng ;
Sheu, Shyh-Shyuan ;
Chen, Frederick T. .
IEEE TRANSACTIONS ON COMPUTERS, 2015, 64 (01) :180-190
[5]  
Chen LR, 2017, DES AUT TEST EUROPE, P19, DOI 10.23919/DATE.2017.7926952
[6]  
Chollet F., 2015, Keras
[7]  
DeBardeleben N., 2014, IEEE WORKSH SIL ERR
[8]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[9]  
Dhiman G, 2009, DES AUT CON, P664
[10]   MEMTI: Optimizing On-Chip Nonvolatile Storage for Visual Multitask Inference at the Edge [J].
Donato, Marco ;
Pentecost, Lillian ;
Brooks, David ;
Wei, Gu-Yeon .
IEEE MICRO, 2019, 39 (06) :73-81