Training of quantized deep neural networks using a magnetic tunnel junction-based synapse

被引:2
作者
Greenberg-Toledo, Tzofnat [1 ]
Perach, Ben [1 ]
Hubara, Itay [1 ,2 ]
Soudry, Daniel [1 ]
Kvatinsky, Shahar [1 ]
机构
[1] Technion Israel Inst Technol, Andrew & Erna Viterbi Fac Elect Engn, IL-3200003 Haifa, Israel
[2] Habana Labs Intel Co, Intel Co, Tel Aviv, Israel
基金
欧洲研究理事会;
关键词
magnetic tunnel junction; memristor; deep neural networks; quantized neural networks; MEMORY;
D O I
10.1088/1361-6641/ac251b
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Quantized neural networks (QNNs) are being actively researched as a solution for the computational complexity and memory intensity of deep neural networks. This has sparked efforts to develop algorithms that support both inference and training with quantized weight and activation values, without sacrificing accuracy. A recent example is the GXNOR framework for stochastic training of ternary and binary neural networks (TNNs and BNNs, respectively). In this paper, we show how magnetic tunnel junction (MTJ) devices can be used to support QNN training. We introduce a novel hardware synapse circuit that uses the MTJ stochastic behaviour to support the quantize update. The proposed circuit enables processing near memory (PNM) of QNN training, which subsequently reduces data movement. We simulated MTJ-based stochastic training of a TNN over the MNIST, SVHN, and CIFAR10 datasets and achieved an accuracy of 98.61% , 93.99% 83.02% , respectively (less than 1% degradation compared to the GXNOR algorithm). We evaluated the synapse array performance potential and showed that the proposed synapse circuit can train TNNs in situ, with 18.3TOPs/W 3TOPs/W for weight update.
引用
收藏
页数:14
相关论文
共 50 条
  • [41] Recurrent neural networks made of magnetic tunnel junctions
    Zheng, Qi
    Zhu, Xiaorui
    Mi, Yuanyuan
    Yuan, Zhe
    Xia, Ke
    AIP ADVANCES, 2020, 10 (02)
  • [42] Deep Energy: Task Driven Training of Deep Neural Networks
    Golts, Alona
    Freedman, Daniel
    Elad, Michael
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2021, 15 (02) : 324 - 338
  • [43] On Calibration of Mixup Training for Deep Neural Networks
    Maronas, Juan
    Ramos, Daniel
    Paredes, Roberto
    STRUCTURAL, SYNTACTIC, AND STATISTICAL PATTERN RECOGNITION, S+SSPR 2020, 2021, 12644 : 67 - 76
  • [44] A novel magnetic tunnel junction structure using the edge of a magnetic film
    Sugiyama, H.
    Inokuchi, T.
    Saito, Y.
    JOURNAL OF MAGNETISM AND MAGNETIC MATERIALS, 2007, 310 (02) : 2003 - 2005
  • [45] Training deep neural networks for wireless sensor networks using loosely and weakly labeled images
    Zhou, Qianwei
    Chen, Yuhang
    Li, Baoqing
    Li, Xiaoxin
    Zhou, Chen
    Huang, Jingchang
    Hu, Haigen
    NEUROCOMPUTING, 2021, 427 : 64 - 73
  • [46] Magnetic anisotropy study for GaMnAs-based magnetic tunnel junction
    Uemura, T
    Sone, T
    Matsuda, K
    Yamamoto, M
    JAPANESE JOURNAL OF APPLIED PHYSICS PART 2-LETTERS & EXPRESS LETTERS, 2005, 44 (42-45): : L1352 - L1354
  • [47] Dynamic Memory Management for GPU-based training of Deep Neural Networks
    Shriram, S. B.
    Garg, Anshuj
    Kulkarni, Purushottam
    2019 IEEE 33RD INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM (IPDPS 2019), 2019, : 200 - 209
  • [48] Supporting the Momentum Training Algorithm Using a Memristor-Based Synapse
    Greenberg-Toledo, Tzofnat
    Mazor, Roee
    Haj-Ali, Ameer
    Kvatinsky, Shahar
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2019, 66 (04) : 1571 - 1583
  • [49] IMPROVEMENTS TO SPEAKER ADAPTIVE TRAINING OF DEEP NEURAL NETWORKS
    Miao, Yajie
    Jiang, Lu
    Zhang, Hao
    Metze, Florian
    2014 IEEE WORKSHOP ON SPOKEN LANGUAGE TECHNOLOGY SLT 2014, 2014, : 165 - 170
  • [50] Skyrmion-Induced Memristive Magnetic Tunnel Junction for Ternary Neural Network
    Pan, Biao
    Zhang, Deming
    Zhang, Xueying
    Wang, Haotian
    Bai, Jinyu
    Yang, Jianlei
    Zhang, Youguang
    Kang, Wang
    Zhao, Weisheng
    IEEE JOURNAL OF THE ELECTRON DEVICES SOCIETY, 2019, 7 (01): : 529 - 533