Neural Network Acceleration and Voice Recognition with a Flash-based In-Memory Computing SoC

被引:6
作者
Zhao, Liang [1 ,3 ]
Gao, Shifan [1 ]
Zhang, Shengbo [2 ]
Qiu, Xiang [2 ]
Yang, Fan [1 ]
Li, Jie [3 ]
Chen, Zezhi [3 ]
Zhao, Yi [1 ,4 ]
机构
[1] Zhejiang Univ, Coll Informat Sci & Elect Engn, Hangzhou, Peoples R China
[2] Flash Bill Semicond Co Ltd, Shanghai, Peoples R China
[3] Hefei Reliance Memory Ltd, Hefei, Peoples R China
[4] Nanhu Acad Elect & Informat Technol, Jiaxing, Peoples R China
来源
2021 IEEE 3RD INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE CIRCUITS AND SYSTEMS (AICAS) | 2021年
关键词
D O I
10.1109/AICAS51828.2021.9458476
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
AI inference based on novel compute-in-memory devices has shown clear advantages in terms of power, speed and storage density, making it a promising candidate for IoT and edge computing applications. In this work, we demonstrate a fully integrated system-on-chip (SoC) design with embedded Flash memories as the neural network accelerator. A series of techniques from device, design and system perspectives are combined to enable efficient AI inference for resource-constrained voice recognition. 7-bit/cell storage capability and self-adaptive write of novel Flash memories are leveraged to achieve state-of-the-art overall performance. Also, model deployment techniques based on transfer learning are explored to significantly improve the accuracy loss during weight data deployment. Integrated in a compact form factor, the whole voice recognition system can achieve >10 TOPS/W energy efficiency and similar to 95% accuracy for realtime keyword spotting applications.
引用
收藏
页数:5
相关论文
共 13 条
[1]  
Burr G.W., 2014, IEEE INT ELECT DEVIC, P1
[2]   TIME:A Training-in-memory Architecture for Memristor-based Deep Neural Networks [J].
Cheng, Ming ;
Xia, Lixue ;
Zhu, Zhenhua ;
Cai, Yi ;
Xie, Yuan ;
Wang, Yu ;
Yang, Huazhong .
PROCEEDINGS OF THE 2017 54TH ACM/EDAC/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2017,
[3]  
Eliasmith, 2019, P 7 ANN NEUR COMP EL
[4]   Programmable Linear RAM: A New Flash Memory-based Memristor for Artificial Synapses and Its Application to Speech Recognition System [J].
Gao, Shifan ;
Yang, Guangjun ;
Qiu, Xiang ;
Yang, Chun ;
Zhang, Cheng ;
Li, Binhan ;
Gao, Chao ;
Jiang, Hong ;
Wang, Zhexian ;
Hu, Jian ;
Xiao, Jun ;
Zhang, Bo ;
Lee, Choonghyun ;
Zhao, Yi ;
Kong, Weiran .
2019 IEEE INTERNATIONAL ELECTRON DEVICES MEETING (IEDM), 2019,
[5]   SpotTune: Transfer Learning through Adaptive Fine-tuning [J].
Guo, Yunhui ;
Shi, Honghui ;
Kumar, Abhishek ;
Grauman, Kristen ;
Rosing, Tajana ;
Feris, Rogerio .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :4800-4809
[6]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[7]   In-memory computing with resistive switching devices [J].
Ielmini, Daniele ;
Wong, H. -S. Philip .
NATURE ELECTRONICS, 2018, 1 (06) :333-343
[8]   Input Voltage Mapping Optimized for Resistive Memory-Based Deep Neural Network Hardware [J].
Kim, Taesu ;
Kim, Hyungjun ;
Kim, Jinseok ;
Kim, Jae-Joon .
IEEE ELECTRON DEVICE LETTERS, 2017, 38 (09) :1228-1231
[9]   RRAM-based 7T1R Nonvolatile SRAM with 2x Reduction in Store Energy and 94x Reduction in Restore Energy for Frequent-Off Instant-On Applications [J].
Lee, Albert ;
Chang, Meng-Fan ;
Lin, Chien-Chen ;
Chen, Chien-Fu ;
Ho, Mon-Shu ;
Kuo, Chia-Chen ;
Tseng, Pei-Ling ;
Sheu, Shyh-Shyuan ;
Ku, Tzu-Kun .
2015 SYMPOSIUM ON VLSI CIRCUITS (VLSI CIRCUITS), 2015,
[10]   Learning to Predict IR Drop with Effective Training for ReRAM-based Neural Network Hardware [J].
Lee, Sugil ;
Jung, Giju ;
Fouda, Mohammed E. ;
Lee, Jongeun ;
Eltawil, Ahmed ;
Kurdahi, Fadi .
PROCEEDINGS OF THE 2020 57TH ACM/EDAC/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2020,