Efficient Error-Tolerant Quantized Neural Network Accelerators

被引:19
作者
Gambardella, Giulio [1 ]
Kappauf, Johannes [1 ]
Blott, Michaela [1 ]
Doehring, Christoph [2 ]
Kumm, Martin [3 ]
Zip, Peter [4 ]
Vissers, Kees [1 ]
机构
[1] Xilinx Res Labs, Dublin, Ireland
[2] Coburg Univ, Coburg, Germany
[3] Fulda Univ Appl Sci, Fulda, Germany
[4] Univ Kassel, Digital Technol Grp, Kassel, Germany
来源
2019 IEEE INTERNATIONAL SYMPOSIUM ON DEFECT AND FAULT TOLERANCE IN VLSI AND NANOTECHNOLOGY SYSTEMS (DFT) | 2019年
关键词
neural networks; safety; automotive; FPGA; quantized neural networks;
D O I
10.1109/dft.2019.8875314
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Neural Networks are currently one of the most widely deployed machine learning algorithms. In particular, Convolutional Neural Networks (CNNs), are gaining popularity and are evaluated for deployment in safety critical applications such as self driving vehicles. Modern CNNs feature enormous memory bandwidth and high computational needs, challenging existing hardware platforms to meet throughput, latency and power requirements. Functional safety and error tolerance need to be considered as additional requirement in safety critical systems. In general, fault tolerant operation can be achieved by adding redundancy to the system, which is further exacerbating the computational demands. Furthermore, the question arises whether pruning and quantization methods for performance scaling turn out to be counterproductive with regards to fail safety requirements. In this work we present a methodology to evaluate the impact of permanent faults affecting Quantized Neural Networks (QNNs) and how to effectively decrease their effects in hardware accelerators. We use FPGA-based hardware accelerated error injection, in order to enable the fast evaluation. A detailed analysis is presented showing that QNNs containing convolutional layers are by far not as robust to faults as commonly believed and can lead to accuracy drops of up to 10%. To circumvent that, we propose two different methods to increase their robustness: 1) selective channel replication which adds significantly less redundancy than used by the common triple modular redundancy and 2) a fault-aware scheduling of processing elements for folded implementations.
引用
收藏
页数:6
相关论文
共 17 条
[1]  
[Anonymous], 2018, FINN-R: An End-to-End Deep-Learning Framework for Fast Exploration of Quantized Neural Networks
[2]  
Bosio A., 2018, ARC 18, P1
[3]  
Courbariaux M., 2016, ARXIV160202830
[4]   Evaluation and Mitigation of Soft-Errors in Neural Network-based Object Detection in Three GPU Architectures [J].
dos Santos, Fernando Fernandes ;
Draghetti, Lucas ;
Weigel, Lucas ;
Carro, Luigi ;
Navaux, Philippe ;
Rech, Paolo .
2017 47TH ANNUAL IEEE/IFIP INTERNATIONAL CONFERENCE ON DEPENDABLE SYSTEMS AND NETWORKS WORKSHOPS (DSN-W 2017), 2017, :169-176
[5]   AI2: Safety and Robustness Certification of Neural Networks with Abstract Interpretation [J].
Gehr, Timon ;
Mirman, Matthew ;
Drachsler-Cohen, Dana ;
Tsankov, Petar ;
Chaudhuri, Swarat ;
Vechev, Martin .
2018 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2018, :3-18
[6]   Using a failure modes, effects and diagnostic analysis (FMEDA) to measure diagnostic coverage in programmable electronic systems [J].
Goble, WM ;
Brombacher, AC .
RELIABILITY ENGINEERING & SYSTEM SAFETY, 1999, 66 (02) :145-148
[7]  
Guo K., 2017, A Survey of FPGA-Based Neural Network Accelerator," dec
[8]  
Howard AG, 2017, ARXIV
[9]  
Jha S, 2018, ARXIV, P1, DOI [DOI 10.15398/JLM.V7I2.214, 10.15398/jlm.v7i2.214]
[10]   Hand gesture recognition based on convolution neural network [J].
Li, Gongfa ;
Tang, Heng ;
Sun, Ying ;
Kong, Jianyi ;
Jiang, Guozhang ;
Jiang, Du ;
Tao, Bo ;
Xu, Shuang ;
Liu, Honghai .
CLUSTER COMPUTING-THE JOURNAL OF NETWORKS SOFTWARE TOOLS AND APPLICATIONS, 2019, 22 (02) :S2719-S2729