In-Memory Computing Architecture for a Convolutional Neural Network Based on Spin Orbit Torque MRAM

被引:1
作者
Huang, Jun-Ying [1 ]
Syu, Jing-Lin [2 ]
Tsou, Yao-Tung [2 ]
Kuo, Sy-Yen [1 ]
Chang, Ching-Ray [3 ]
机构
[1] Natl Taiwan Univ, Dept Elect Engn, Taipei 106, Taiwan
[2] Feng Chia Univ, Dept Commun Engn, Taichung 407, Taiwan
[3] Chung Yuan Christian Univ, Quantum Informat Ctr, Taoyuan 320, Taiwan
关键词
convolution neural network; computing in memory; processing in memory; distributed arithmetic; MRAM; SOT-MRAM; ENERGY;
D O I
10.3390/electronics11081245
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Recently, numerous studies have investigated computing in-memory (CIM) architectures for neural networks to overcome memory bottlenecks. Because of its low delay, high energy efficiency, and low volatility, spin-orbit torque magnetic random access memory (SOT-MRAM) has received substantial attention. However, previous studies used calculation circuits to support complex calculations, leading to substantial energy consumption. Therefore, our research proposes a new CIM architecture with small peripheral circuits; this architecture achieved higher performance relative to other CIM architectures when processing convolution neural networks (CNNs). We included a distributed arithmetic (DA) algorithm to improve the efficiency of the CIM calculation method by reducing the excessive read/write times and execution steps of CIM-based CNN calculation circuits. Furthermore, our method also uses SOT-MRAM to increase the calculation speed and reduce power consumption. Compared with CIM-based CNN arithmetic circuits in previous studies, our method can achieve shorter clock periods and reduce read times by up to 43.3% without the need for additional circuits.
引用
收藏
页数:17
相关论文
共 19 条
[1]  
Albawi S, 2017, I C ENG TECHNOL
[2]  
Alwani M, 2016, INT SYMP MICROARCH
[3]  
Angizi S, 2018, ASIA S PACIF DES AUT, P111, DOI 10.1109/ASPDAC.2018.8297291
[4]  
Binkert Nathan, 2011, Computer Architecture News, V39, P1, DOI 10.1145/2024716.2024718
[5]   Area-Efficient Distributed Arithmetic Optimization via Heuristic Decomposition and In-Memroy Computing [J].
Chen, Jian ;
Zhao, Wenfeng ;
Ha, Yajun .
2019 IEEE 13TH INTERNATIONAL CONFERENCE ON ASIC (ASICON), 2019,
[6]   PRIME: A Novel Processing-in-memory Architecture for Neural Network Computation in ReRAM-based Main Memory [J].
Chi, Ping ;
Li, Shuangchen ;
Xu, Cong ;
Zhang, Tao ;
Zhao, Jishen ;
Liu, Yongpan ;
Wang, Yu ;
Xie, Yuan .
2016 ACM/IEEE 43RD ANNUAL INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE (ISCA), 2016, :27-39
[7]   DrAcc: a DRAM based Accelerator for Accurate CNN Inference [J].
Deng, Quan ;
Jiang, Lei ;
Zhang, Youtao ;
Zhang, Minxuan ;
Yang, Jun .
2018 55TH ACM/ESDA/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2018,
[8]   NVSim: A Circuit-Level Performance, Energy, and Area Model for Emerging Nonvolatile Memory [J].
Dong, Xiangyu ;
Xu, Cong ;
Xie, Yuan ;
Jouppi, Norman P. .
IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2012, 31 (07) :994-1007
[9]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[10]   Compact Model for Spin-Orbit Magnetic Tunnel Junctions [J].
Kazemi, Mohammad ;
Rowlands, Graham E. ;
Ipek, Engin ;
Buhrman, Robert A. ;
Friedman, Eby G. .
IEEE TRANSACTIONS ON ELECTRON DEVICES, 2016, 63 (02) :848-855