Design of Computing-in-Memory (CIM) with Vertical Split-Gate Flash Memory for Deep Neural Network (DNN) Inference Accelerator

被引:12
作者
Lue, Hang-Ting [1 ]
Hu, Han-Wen [1 ]
Hsu, Tzu-Hsuan [1 ]
Hsu, Po-Kai [1 ]
Wang, Keh-Chung [1 ]
Lu, Chih-Yuan [1 ]
机构
[1] Macronix Int Co Ltd, 16 Li Hsin Rd,Hsinchu Sci Pk, Hsinchu, Taiwan
来源
2021 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS) | 2021年
关键词
D O I
10.1109/ISCAS51556.2021.9401723
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Computing-In-Memory (CIM) using Flash memory is a potential solution to support a heavy-weight DNN inference accelerator for edge computing applications. Flash memory provides the best high-density and low-cost non-volatile memory solution to store the weights, while CIM functions of Flash memory can compute AI neural network calculations inside the memory chip. Our analysis indicates that Flash CIM can save data movements by similar to 85% as compared with the conventional Von-Neumann architecture. In this work, we propose a detail device and design co-optimizations to realize Flash CIM, using a novel vertical split-gate Flash device. Our device supports low-voltage (<1V) read at WL's and BL's, tight and tunable cell current (Icell) ranging from 150nA to 1.5uA, extremely large Icell ON/OFF ratio similar to 7 orders, small RTN noise and negligible read disturb to provide a high-performance and highly-reliable CIM solution.
引用
收藏
页数:4
相关论文
共 5 条
[1]  
Chang M. F., VLSI
[2]  
Chen W. H., ISSCC, P494
[3]  
Lue H. T., IEDM, pS38
[4]  
Lue H. T., VLSI, P347
[5]  
Xue C. X., ISSCC, P388