Saving Memory Space in Deep Neural Networks by Recomputing: A Survey

被引:0
|
作者
Ulidowski, Irek [1 ,2 ]
机构
[1] Univ Leicester, Sch Comp & Math Sci, Leicester, Leics, England
[2] AGH Univ Sci & Technol, Dept Appl Informat, Krakow, Poland
来源
REVERSIBLE COMPUTATION, RC 2023 | 2023年 / 13960卷
关键词
Deep Neural Networks; recomputing activations;
D O I
10.1007/978-3-031-38100-3_7
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Training a multilayered neural network involves execution of the network on the training data, followed by calculating the error between the predicted and actual output, and then performing backpropagation to update the network's weights in order to minimise the overall error. This process is repeated many times, with the network updating its weights until it produces the desired output with a satisfactory level of accuracy. It requires storage in memory of activation and gradient data for each layer during each training run of the network. This paper surveys the main approaches to recomputing the needed activation and gradient data instead of storing it in memory. We discuss how these approaches relate to reversible computation techniques.
引用
收藏
页码:89 / 105
页数:17
相关论文
共 50 条
  • [41] Triangular Trade-off between Robustness, Accuracy, and Fairness in Deep Neural Networks: A Survey
    Li, Jingyang
    Li, Guoqiang
    ACM COMPUTING SURVEYS, 2025, 57 (06)
  • [42] Deep neural networks for automatic speech processing: a survey from large corpora to limited data
    Roger, Vincent
    Farinas, Jerome
    Pinquier, Julien
    EURASIP JOURNAL ON AUDIO SPEECH AND MUSIC PROCESSING, 2022, 2022 (01)
  • [43] Dynamic Slicing for Deep Neural Networks
    Zhang, Ziqi
    Li, Yuanchun
    Guo, Yao
    Chen, Xiangqun
    Liu, Yunxin
    PROCEEDINGS OF THE 28TH ACM JOINT MEETING ON EUROPEAN SOFTWARE ENGINEERING CONFERENCE AND SYMPOSIUM ON THE FOUNDATIONS OF SOFTWARE ENGINEERING (ESEC/FSE '20), 2020, : 838 - 850
  • [44] A Representer Theorem for Deep Neural Networks
    Unser, Michael
    JOURNAL OF MACHINE LEARNING RESEARCH, 2019, 20
  • [45] Conceptual alignment deep neural networks
    Dai, Yinglong
    Wang, Guojun
    Li, Kuan-Ching
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2018, 34 (03) : 1631 - 1642
  • [46] The Representation of Speech in Deep Neural Networks
    Scharenborg, Odette
    van der Gouw, Nikki
    Larson, Martha
    Marchiori, Elena
    MULTIMEDIA MODELING, MMM 2019, PT II, 2019, 11296 : 194 - 205
  • [47] Temporal Alignment for Deep Neural Networks
    Lin, Payton
    Lyu, Dau-Cheng
    Chang, Yun-Fan
    Tsao, Yu
    2015 IEEE GLOBAL CONFERENCE ON SIGNAL AND INFORMATION PROCESSING (GLOBALSIP), 2015, : 108 - 112
  • [48] Embedding Watermarks into Deep Neural Networks
    Uchida, Yusuke
    Nagai, Yuki
    Sakazawa, Shigeyuki
    Satoh, Shin'ichi
    PROCEEDINGS OF THE 2017 ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL (ICMR'17), 2017, : 274 - 282
  • [49] Subaging in underparametrized deep neural networks
    Herrera Segura, Carolina
    Montoya, Edison
    Tapias, Diego
    MACHINE LEARNING-SCIENCE AND TECHNOLOGY, 2022, 3 (03):
  • [50] Visual Genealogy of Deep Neural Networks
    Wang, Qianwen
    Yuan, Jun
    Chen, Shuxin
    Su, Hang
    Qu, Huamin
    Liu, Shixia
    IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2020, 26 (11) : 3340 - 3352