A Ferroelectric FET-Based Processing-in-Memory Architecture for DNN Acceleration

被引:56
作者
Long, Yun [1 ]
Kim, Daehyun [1 ]
Lee, Edward [1 ]
Saha, Priyabrata [1 ]
Mudassar, Burhan Ahmad [1 ]
She, Xueyuan [1 ]
Khan, Asif Islam [1 ]
Mukhopadhyay, Saibal [1 ]
机构
[1] Georgia Inst Technol, Sch Elect & Comp Engn, Atlanta, GA 30332 USA
来源
IEEE JOURNAL ON EXPLORATORY SOLID-STATE COMPUTATIONAL DEVICES AND CIRCUITS | 2019年 / 5卷 / 02期
基金
美国国家科学基金会;
关键词
Virtual machine monitors; Engines; Clocks; Transistors; Logic gates; Computer architecture; Random access memory; Deep neural network (DNN); ferroelectric FET (FeFET); processing-in-memory (PIM);
D O I
10.1109/JXCDC.2019.2923745
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
This paper presents a ferroelectric FET (FeFET)-based processing-in-memory (PIM) architecture to accelerate the inference of deep neural networks (DNNs). We propose a digital in-memory vector-matrix multiplication (VMM) engine design utilizing the FeFET crossbar to enable bit-parallel computation and eliminate analog-to-digital conversion in prior mixed-signal PIM designs. A dedicated hierarchical network-on-chip (H-NoC) is developed for input broadcasting and on-the-fly partial results processing, reducing the data transmission volume and latency. Simulations in 28-nm CMOS technology show 115xand 6.3xhigher computing efficiency (GOPs/W) over desktop GPU (NvidiaGTX1080Ti) and resistive random access memory (ReRAM)-based design, respectively.
引用
收藏
页码:113 / 122
页数:10
相关论文
共 25 条
[1]  
[Anonymous], ACM SIGARCH COMPUTER
[2]  
[Anonymous], 2017, ABS170404861 CORR
[3]  
[Anonymous], THESIS
[4]  
[Anonymous], 2015, P 42 ANN INT S COMP
[5]  
[Anonymous], 2018, P INT C COMP AID DES
[6]  
Aziz A, 2018, DES AUT TEST EUROPE, P1289, DOI 10.23919/DATE.2018.8342213
[7]  
Chen XM, 2018, DES AUT TEST EUROPE, P1205, DOI 10.23919/DATE.2018.8342199
[8]   DaDianNao: A Machine-Learning Supercomputer [J].
Chen, Yunji ;
Luo, Tao ;
Liu, Shaoli ;
Zhang, Shijin ;
He, Liqiang ;
Wang, Jia ;
Li, Ling ;
Chen, Tianshi ;
Xu, Zhiwei ;
Sun, Ninghui ;
Temam, Olivier .
2014 47TH ANNUAL IEEE/ACM INTERNATIONAL SYMPOSIUM ON MICROARCHITECTURE (MICRO), 2014, :609-622
[9]   PRIME: A Novel Processing-in-memory Architecture for Neural Network Computation in ReRAM-based Main Memory [J].
Chi, Ping ;
Li, Shuangchen ;
Xu, Cong ;
Zhang, Tao ;
Zhao, Jishen ;
Liu, Yongpan ;
Wang, Yu ;
Xie, Yuan .
2016 ACM/IEEE 43RD ANNUAL INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE (ISCA), 2016, :27-39
[10]   Neural Cache: Bit-Serial In-Cache Acceleration of Deep Neural Networks [J].
Eckert, Charles ;
Wang, Xiaowei ;
Wang, Jingcheng ;
Subramaniyan, Arun ;
Iyer, Ravi ;
Sylvester, Dennis ;
Blaauw, David ;
Das, Reetuparna .
2018 ACM/IEEE 45TH ANNUAL INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE (ISCA), 2018, :383-396