Mixed-precision architecture based on computational memory for training deep neural networks

被引:63
|
作者
Nandakumar, S. R. [1 ,2 ]
Le Gallo, Manuel [1 ]
Boybat, Irem [1 ,3 ]
Rajendran, Bipin [2 ]
Sebastian, Abu [1 ]
Eleftheriou, Evangelos [1 ]
机构
[1] IBM Res Zurich, CH-8803 Ruschlikon, Switzerland
[2] New Jersey Inst Technol, Newark, NJ 07102 USA
[3] Ecole Polytech Fed Lausanne, CH-1015 Lausanne, Switzerland
关键词
Deep learning; In-memory computing; Mixed-precision computing; Phase-change memory;
D O I
10.1109/ISCAS.2018.8351656
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Deep neural networks (DNN) have revolutionized the field of machine learning by providing unprecedented human-like performance in solving many real-world problems such as image or speech recognition. Training of large DNNs, however, is a computationally intensive task, and this necessitates the development of novel computing architectures targeting this application. A computational memory unit where resistive memory devices are organized in crossbar arrays can be used to store the synaptic weights in their conductance states. The expensive multiply accumulate operations can be performed in place using Kirchhoff's circuit laws in a non-von Neumann manner. However, a key challenge remains the inability to alter the conductance states of the devices in a reliable manner during the weight update process. We propose a mixed-precision architecture that combines a computational memory unit storing the synaptic weights with a digital processing unit and an additional memory unit that stores the accumulated weight updates in high precision. The new architecture delivers classification accuracies comparable to those of floating-point implementations without being constrained by challenges associated with the non-ideal weight update characteristics of emerging resistive memories. The computational memory unit in a two layer neural network realized using non-linear stochastic models of phase-change memory achieves a test accuracy of 97.40% in the MNIST digit classification problem.
引用
收藏
页数:5
相关论文
共 50 条
  • [1] Mixed-Precision Deep Learning Based on Computational Memory
    Nandakumar, S. R.
    Le Gallo, Manuel
    Piveteau, Christophe
    Joshi, Vinay
    Mariani, Giovanni
    Boybat, Irem
    Karunaratne, Geethan
    Khaddam-Aljameh, Riduan
    Egger, Urs
    Petropoulos, Anastasios
    Antonakopoulos, Theodore
    Rajendran, Bipin
    Sebastian, Abu
    Eleftheriou, Evangelos
    FRONTIERS IN NEUROSCIENCE, 2020, 14
  • [2] Hardware for Quantized Mixed-Precision Deep Neural Networks
    Rios, Andres
    Nava, Patricia
    PROCEEDINGS OF THE 2022 15TH IEEE DALLAS CIRCUITS AND SYSTEMS CONFERENCE (DCAS 2022), 2022,
  • [3] MINT: Mixed-precision RRAM-based IN-memory Training Architecture
    Jiang, Hongwu
    Huang, Shanshi
    Peng, Xiaochen
    Yu, Shimeng
    2020 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2020,
  • [4] Exploration of Automatic Mixed-Precision Search for Deep Neural Networks
    Guo, Xuyang
    Huang, Yuanjun
    Cheng, Hsin-pai
    Li, Bing
    Wen, Wei
    Ma, Siyuan
    Li, Hai
    Chen, Yiran
    2019 IEEE INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE CIRCUITS AND SYSTEMS (AICAS 2019), 2019, : 276 - 278
  • [5] Mixed-precision quantization-aware training for photonic neural networks
    Kirtas, Manos
    Passalis, Nikolaos
    Oikonomou, Athina
    Moralis-Pegios, Miltos
    Giamougiannis, George
    Tsakyridis, Apostolos
    Mourgias-Alexandris, George
    Pleros, Nikolaos
    Tefas, Anastasios
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (29): : 21361 - 21379
  • [6] Mixed-precision quantization-aware training for photonic neural networks
    Manos Kirtas
    Nikolaos Passalis
    Athina Oikonomou
    Miltos Moralis-Pegios
    George Giamougiannis
    Apostolos Tsakyridis
    George Mourgias-Alexandris
    Nikolaos Pleros
    Anastasios Tefas
    Neural Computing and Applications, 2023, 35 : 21361 - 21379
  • [7] Hessian-based mixed-precision quantization with transition aware training for neural networks
    Huang, Zhiyong
    Han, Xiao
    Yu, Zhi
    Zhao, Yunlan
    Hou, Mingyang
    Hu, Shengdong
    NEURAL NETWORKS, 2025, 182
  • [8] EVOLUTIONARY QUANTIZATION OF NEURAL NETWORKS WITH MIXED-PRECISION
    Liu, Zhenhua
    Zhang, Xinfeng
    Wang, Shanshe
    Ma, Siwei
    Gao, Wen
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 2785 - 2789
  • [9] Evaluating the Impact of Mixed-Precision on Fault Propagation for Deep Neural Networks on GPUs
    Dos Santos, Fernando Fernandes
    Rech, Paolo
    Kritikakou, Angeliki
    Sentieys, Olivier
    2022 IEEE COMPUTER SOCIETY ANNUAL SYMPOSIUM ON VLSI (ISVLSI 2022), 2022, : 327 - 327
  • [10] Computational memory-based inference and training of deep neural networks
    Sebastian, A.
    Boybat, I.
    Dazzi, M.
    Giannopoulos, I.
    Jonnalagadda, V.
    Joshi, V.
    Karunaratne, G.
    Kersting, B.
    Khaddam-Aljameh, R.
    Nandakumar, S. R.
    Petropoulos, A.
    Piveteau, C.
    Antonakopoulos, T.
    Rajendran, B.
    Le Gallo, M.
    Eleftheriou, E.
    2019 SYMPOSIUM ON VLSI CIRCUITS, 2019, : T168 - T169