Saving Memory Space in Deep Neural Networks by Recomputing: A Survey

被引:0
|
作者
Ulidowski, Irek [1 ,2 ]
机构
[1] Univ Leicester, Sch Comp & Math Sci, Leicester, Leics, England
[2] AGH Univ Sci & Technol, Dept Appl Informat, Krakow, Poland
来源
REVERSIBLE COMPUTATION, RC 2023 | 2023年 / 13960卷
关键词
Deep Neural Networks; recomputing activations;
D O I
10.1007/978-3-031-38100-3_7
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Training a multilayered neural network involves execution of the network on the training data, followed by calculating the error between the predicted and actual output, and then performing backpropagation to update the network's weights in order to minimise the overall error. This process is repeated many times, with the network updating its weights until it produces the desired output with a satisfactory level of accuracy. It requires storage in memory of activation and gradient data for each layer during each training run of the network. This paper surveys the main approaches to recomputing the needed activation and gradient data instead of storing it in memory. We discuss how these approaches relate to reversible computation techniques.
引用
收藏
页码:89 / 105
页数:17
相关论文
共 50 条
  • [31] Orthogonal Deep Neural Networks
    Li, Shuai
    Jia, Kui
    Wen, Yuxin
    Liu, Tongliang
    Tao, Dacheng
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2021, 43 (04) : 1352 - 1368
  • [32] A survey on deep neural networks for human action recognition in RGB image and depth image
    Wang, Hongyu
    ENERGY SCIENCE AND APPLIED TECHNOLOGY (ESAT 2016), 2016, : 697 - 703
  • [33] Deep Neural Networks for Spatial-Temporal Cyber-Physical Systems: A Survey
    Musa, Abubakar Ahmad
    Hussaini, Adamu
    Liao, Weixian
    Liang, Fan
    Yu, Wei
    FUTURE INTERNET, 2023, 15 (06)
  • [34] On the Opacity of Deep Neural Networks
    Sogaard, Anders
    CANADIAN JOURNAL OF PHILOSOPHY, 2024, : 224 - 239
  • [35] A Survey on Side-Channel-based Reverse Engineering Attacks on Deep Neural Networks
    Liu, Yuntao
    Zuzak, Michael
    Xing, Daniel
    McDaniel, Isaac
    Mittu, Priya
    Ozbay, Olsan
    Akib, Abir
    Srivastava, Ankur
    2022 IEEE INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE CIRCUITS AND SYSTEMS (AICAS 2022): INTELLIGENT TECHNOLOGY IN THE POST-PANDEMIC ERA, 2022, : 312 - 315
  • [36] Convolutional long short term memory deep neural networks for image sequence prediction
    Balderas, David
    Ponce, Pedro
    Molina, Arturo
    EXPERT SYSTEMS WITH APPLICATIONS, 2019, 122 : 152 - 162
  • [37] A Design Space Exploration Framework for Deployment of Resource-Constrained Deep Neural Networks
    Zhang, Yan
    Pan, Lei
    Berkowitz, Phillip
    Lee, Mun Wai
    Riggan, Benjamin
    Bhattacharyya, Shuvra S.
    REAL-TIME IMAGE PROCESSING AND DEEP LEARNING 2024, 2024, 13034
  • [38] Memory-Based Pruning of Deep Neural Networks for IoT Devices Applied to Flood Detection
    Fernandes Junior, Francisco Erivaldo
    Nonato, Luis Gustavo
    Ranieri, Caetano Mazzoni
    Ueyama, Jo
    SENSORS, 2021, 21 (22)
  • [39] Ternary Compute-Enabled Memory using Ferroelectric Transistors for Accelerating Deep Neural Networks
    Thirumala, Sandeep Krishna
    Jain, Shubham
    Gupta, Sumeet Kumar
    Raghunathan, Anand
    PROCEEDINGS OF THE 2020 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE 2020), 2020, : 31 - 36
  • [40] Deep neural networks for automatic speech processing: a survey from large corpora to limited data
    Vincent Roger
    Jérôme Farinas
    Julien Pinquier
    EURASIP Journal on Audio, Speech, and Music Processing, 2022