An Energy-Efficient High CSNR XNOR and Accumulation Scheme For BNN

被引:7
作者
Kushwaha, Dinesh [1 ]
Joshi, Ashish [2 ]
Kumar, Chaudhry Indra [3 ]
Gupta, Neha [1 ]
Miryala, Sandeep [4 ]
Joshi, Rajiv, V [5 ]
Dasgupta, Sudeb [1 ]
Bulusu, Anand [1 ]
机构
[1] Indian Inst Technol Roorkee, Dept Elect & Commun Engn, Roorkee 247667, Uttar Pradesh, India
[2] Intel Technol India Pvt Ltd, IP Engn Grp IPG, Bengaluru 560103, India
[3] Delhi Technol Univ, Dept Elect Engn, New Delhi 110042, India
[4] Brookhaven Natl Lab, Instrumentat Div, Upton, NY 11973 USA
[5] Thomas J Watson Res Ctr, Yorktown Hts, NY 10598 USA
关键词
Accuracy; accumulation; artificial intelligence (AI); compute signal margin (CSM); compute signal to noise ratio (CSNR); energy-efficiency; latency; neuron; SRAM; COMPUTING SRAM MACRO; IN-MEMORY MACRO; CNN ACCELERATOR;
D O I
10.1109/TCSII.2022.3149818
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In this brief, we present an energy-efficient and high compute signal-to-noise ratio (CSNR) XNOR and accumulation (XAC) scheme for binary neural networks (BNNs). Transmission gates achieve a large compute signal margin (CSM) and high CSNR for accurate XAC operation. The 10T1C XNOR SRAM bit-cell performs the in-memory XAC operation without pre-charging the larger bitline capacitances and significantly reducing energy consumption per XAC operation. The validation of the proposed XAC scheme is done through the post-layout simulations in 65nm CMOS technology with V-DD of 1 V. The achieved 1 ns of latency and 2.36 fJ of energy consumption per XAC operation are (7.2 x , 7.2 x ) and (2 x , 1.31 x ) lower than state-of-the-art digital and analog compute in-memory (CIM) XAC schemes respectively. The proposed XAC design achieves 8.6 x improvement in figure-of-merit (FoM), over prior state-of-the-art. Moreover, (sigma/mu) average of 0.2% from Monte Carlo simulations show that proposed XAC scheme is robust against systematic mismatch and process variations.
引用
收藏
页码:2311 / 2315
页数:5
相关论文
共 17 条
[1]   Xcel-RAM: Accelerating Binary Neural Networks in High-Throughput SRAM Compute Arrays [J].
Agrawal, Amogh ;
Jaiswal, Akhilesh ;
Roy, Deboleena ;
Han, Bing ;
Srinivasan, Gopalakrishnan ;
Ankit, Aayush ;
Roy, Kaushik .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2019, 66 (08) :3064-3076
[2]  
Biswas A, 2018, ISSCC DIG TECH PAP I, P488, DOI 10.1109/ISSCC.2018.8310397
[3]   C3SRAM: An In-Memory-Computing SRAM Macro Based on Robust Capacitive Coupling Computing Mechanism [J].
Jiang, Zhewei ;
Yin, Shihui ;
Seo, Jae-Sun ;
Seok, Mingoo .
IEEE JOURNAL OF SOLID-STATE CIRCUITS, 2020, 55 (07) :1888-1897
[4]   An In-Memory VLSI Architecture for Convolutional Neural Networks [J].
Kang, Mingu ;
Lim, Sungmin ;
Gonugondla, Sujan ;
Shanbhag, Naresh R. .
IEEE JOURNAL ON EMERGING AND SELECTED TOPICS IN CIRCUITS AND SYSTEMS, 2018, 8 (03) :494-505
[5]   Colonnade: A Reconfigurable SRAM-Based Digital Bit-Serial Compute-In-Memory Macro for Processing Neural Networks [J].
Kim, Hyunjoon ;
Yoo, Taegeun ;
Kim, Tony Tae-Hyoung ;
Kim, Bongjin .
IEEE JOURNAL OF SOLID-STATE CIRCUITS, 2021, 56 (07) :2221-2233
[6]  
Kushwaha D., 2020, P IBM IEEE CAS EDS A
[7]  
Liu SW, 2020, 2020 2ND IEEE INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE CIRCUITS AND SYSTEMS (AICAS 2020), P21, DOI 10.1109/AICAS48895.2020.9073848
[8]   XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks [J].
Rastegari, Mohammad ;
Ordonez, Vicente ;
Redmon, Joseph ;
Farhadi, Ali .
COMPUTER VISION - ECCV 2016, PT IV, 2016, 9908 :525-542
[9]  
Si X, 2020, ISSCC DIG TECH PAP I, P246, DOI [10.1109/isscc19947.2020.9062995, 10.1109/ISSCC19947.2020.9062995]
[10]   A Dual-Split 6T SRAM-Based Computing-in-Memory Unit-Macro With Fully Parallel Product-Sum Operation for Binarized DNN Edge Processors [J].
Si, Xin ;
Khwa, Win-San ;
Chen, Jia-Jing ;
Li, Jia-Fang ;
Sun, Xiaoyu ;
Liu, Rui ;
Yu, Shimeng ;
Yamauchi, Hiroyuki ;
Li, Qiang ;
Chang, Meng-Fan .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2019, 66 (11) :4172-4185