Saving Memory Space in Deep Neural Networks by Recomputing: A Survey

被引:0
|
作者
Ulidowski, Irek [1 ,2 ]
机构
[1] Univ Leicester, Sch Comp & Math Sci, Leicester, Leics, England
[2] AGH Univ Sci & Technol, Dept Appl Informat, Krakow, Poland
来源
REVERSIBLE COMPUTATION, RC 2023 | 2023年 / 13960卷
关键词
Deep Neural Networks; recomputing activations;
D O I
10.1007/978-3-031-38100-3_7
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Training a multilayered neural network involves execution of the network on the training data, followed by calculating the error between the predicted and actual output, and then performing backpropagation to update the network's weights in order to minimise the overall error. This process is repeated many times, with the network updating its weights until it produces the desired output with a satisfactory level of accuracy. It requires storage in memory of activation and gradient data for each layer during each training run of the network. This paper surveys the main approaches to recomputing the needed activation and gradient data instead of storing it in memory. We discuss how these approaches relate to reversible computation techniques.
引用
收藏
页码:89 / 105
页数:17
相关论文
共 50 条
  • [21] Dynamic Memory Management for GPU-based training of Deep Neural Networks
    Shriram, S. B.
    Garg, Anshuj
    Kulkarni, Purushottam
    2019 IEEE 33RD INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM (IPDPS 2019), 2019, : 200 - 209
  • [22] Impact of On-chip Interconnect on In-memory Acceleration of Deep Neural Networks
    Krishnan, Gokul
    Mandal, Sumit K.
    Chakrabarti, Chaitali
    Seo, Jae-Sun
    Ogras, Umit Y.
    Cao, Yu
    ACM JOURNAL ON EMERGING TECHNOLOGIES IN COMPUTING SYSTEMS, 2022, 18 (02)
  • [23] Survey of deployment locations and underlying hardware architectures for contemporary deep neural networks
    Kotlar, Milos
    Bojic, Dragan
    Punt, Marija
    Milutinovic, Veljko
    INTERNATIONAL JOURNAL OF DISTRIBUTED SENSOR NETWORKS, 2019, 15 (08)
  • [24] Latent Space-Based Backdoor Attacks Against Deep Neural Networks
    Kristanto, Adrian
    Wang, Shuo
    Rudolph, Carsten
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [25] An Updated Survey of Efficient Hardware Architectures for Accelerating Deep Convolutional Neural Networks
    Capra, Maurizio
    Bussolino, Beatrice
    Marchisio, Alberto
    Shafique, Muhammad
    Masera, Guido
    Martina, Maurizio
    FUTURE INTERNET, 2020, 12 (07):
  • [26] Noise tolerant ternary weight deep neural networks for analog in-memory inference
    Doevenspeck, Jonas
    Vrancx, Peter
    Laubeuf, Nathan
    Mallik, Arindam
    Debacker, Peter
    Verkest, Diederik
    Lauwereins, Rudy
    Dehaene, Wim
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [27] Tracking with Deep Neural Networks
    Kucharczyk, Marcin
    Wolter, Marcin
    PHOTONICS APPLICATIONS IN ASTRONOMY, COMMUNICATIONS, INDUSTRY, AND HIGH-ENERGY PHYSICS EXPERIMENTS 2019, 2019, 11176
  • [28] Built-In Functional Testing of Analog In-Memory Accelerators for Deep Neural Networks
    Mishra, Abhishek Kumar
    Das, Anup Kumar
    Kandasamy, Nagarajan
    ELECTRONICS, 2022, 11 (16)
  • [29] FPGA BASED IMPLEMENTATION OF DEEP NEURAL NETWORKS USING ON-CHIP MEMORY ONLY
    Park, Jinhwan
    Sung, Wonyong
    2016 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING PROCEEDINGS, 2016, : 1011 - 1015
  • [30] Tweaking Deep Neural Networks
    Kim, Jinwook
    Yoon, Heeyong
    Kim, Min-Soo
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (09) : 5715 - 5728