Accuracy improvement in Ag:a-Si memristive synaptic device-based neural network through Adadelta learning method on handwritten-digit recognition

被引:3
作者
Yilmaz, Yildiran [1 ]
机构
[1] Recep Tayyip Erdogan Univ, Dept Comp Engn, TR-53100 Fener, Rize, Turkiye
关键词
Embedded machine learning; Optimisation methods; Neural network; Adadelta; Synaptic device; MEMORY; ANALOG;
D O I
10.1007/s00521-023-08995-y
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Traditional computing architecture (Von Neumann) that requires data transfer between the off-chip memory and processor consumes a large amount of energy when running machine learning (ML) models. Memristive synaptic devices are employed to eliminate this inevitable inefficiency in energy while solving cognitive tasks. However, the performances of energy-efficient neuromorphic systems, which are expected to provide promising results, need to be enhanced in terms of accuracy and test error rates for classification applications. Improving accuracy in such ML models depends on the optimal learning parameter changes from a device to algorithm-level optimisation. To do this, this paper considers the Adadelta, an adaptive learning rate technique, to achieve accurate results by reducing the losses and compares the accuracy, test error rates, and energy consumption of stochastic gradient descent (SGD), Adagrad and Adadelta optimisation methods integrated into the Ag:a-Si synaptic device neural network model. The experimental results demonstrated that Adadelta enhanced the accuracy of the hardware-based neural network model by up to 4.32% when compared to the Adagrad method. The Adadelta method achieved the best accuracy rate of 94%, while DGD and SGD provided an accuracy rate of 68.11 and 75.37%, respectively. These results show that it is vital to select a proper optimisation method to enhance performance, particularly the accuracy and test error rates of the neuro-inspired nano-synaptic device-based neural network models.
引用
收藏
页码:23943 / 23958
页数:16
相关论文
共 48 条
  • [1] A Training-Efficient Hybrid-Structured Deep Neural Network With Reconfigurable Memristive Synapses
    Bai, Kangjun
    An, Qiyuan
    Liu, Lingjia
    Yi, Yang
    [J]. IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS, 2020, 28 (01) : 62 - 75
  • [2] A Survey of Handwritten Character Recognition with MNIST and EMNIST
    Baldominos, Alejandro
    Saez, Yago
    Isasi, Pedro
    [J]. APPLIED SCIENCES-BASEL, 2019, 9 (15):
  • [3] Handwritten Digit Recognition of MNIST dataset using Deep Learning state-of-the-art Artificial Neural Network (ANN) and Convolutional Neural Network (CNN)
    Beohar, Drishti
    Rasool, Akhtar
    [J]. 2021 INTERNATIONAL CONFERENCE ON EMERGING SMART COMPUTING AND INFORMATICS (ESCI), 2021, : 542 - 548
  • [4] Bottou Leon, 2012, Neural Networks: Tricks of the Trade. Second Edition: LNCS 7700, P421, DOI 10.1007/978-3-642-35289-8_25
  • [5] Experimental Demonstration and Tolerancing of a Large-Scale Neural Network (165 000 Synapses) Using Phase-Change Memory as the Synaptic Weight Element
    Burr, Geoffrey W.
    Shelby, Robert M.
    Sidler, Severin
    di Nolfo, Carmelo
    Jang, Junwoo
    Boybat, Irem
    Shenoy, Rohit S.
    Narayanan, Pritish
    Virwani, Kumar
    Giacometti, Emanuele U.
    Kuerdi, Bulent N.
    Hwang, Hyunsang
    [J]. IEEE TRANSACTIONS ON ELECTRON DEVICES, 2015, 62 (11) : 3498 - 3507
  • [6] Technological Benchmark of Analog Synaptic Devices for Neuroinspired Architectures
    Chen, Pai-Yu
    Yu, Shimeng
    [J]. IEEE DESIGN & TEST, 2019, 36 (03) : 31 - 38
  • [7] NeuroSim: A Circuit-Level Macro Model for Benchmarking Neuro-Inspired Architectures in Online Learning
    Chen, Pai-Yu
    Peng, Xiaochen
    Yu, Shimeng
    [J]. IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2018, 37 (12) : 3067 - 3080
  • [8] Zeiler MD, 2012, Arxiv, DOI arXiv:1212.5701
  • [9] KIRCHHOFF CIRCULATION LAW APPLIED TO MULTI-LOOP KINEMATIC CHAINS
    DAVIES, TH
    [J]. MECHANISM AND MACHINE THEORY, 1981, 16 (03) : 171 - 183
  • [10] Dean J., 2012, ADV NEURAL INFORM PR, V25, P1