Facto-CNN: Memory-Efficient CNN Training with Low-rank Tensor Factorization and Lossy Tensor Compression

被引:0
|
作者
Lee, Seungtae [1 ]
Ko, Jonghwan [1 ]
Hong, Seokin [1 ]
机构
[1] Sungkyunkwan Univ, Suwon, South Korea
基金
新加坡国家研究基金会;
关键词
CNN; Training Optimization; Compression; Tensor Factorization;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Convolutional neural networks (CNNs) are becoming deeper and wider to achieve higher accuracy and lower loss, significantly expanding the computational resources. Especially, training CNN models extensively consumes memory mainly due to storing intermediate feature maps generated in the forward-propagation for calculating the gradient in the back-propagation. The memory usage of the CNN model training escalates with the increase in batch size and the complexity of the model. Therefore, a lightweight training method is essential, especially when the computational resources are limited. In this paper, we propose a CNN training mechanism called Facto-CNN, leveraging low-rank tensor factorization and lossy tensor compression to reduce the memory usage required in training the CNN models. Facto-CNN factorizes the weight tensors of convolutional and fully-connected layers and then only updates one of the factorized tensors for each layer, dramatically reducing the feature map size stored in the memory. To further reduce memory consumption, Facto-CNN compresses the feature maps with a simple lossy compression technique that exploits the value similarity in the feature maps. Our experimental evaluation demonstrates that Facto-CNN reduces the memory usage for storing the feature maps by 68-93% with a trivial accuracy degradation when training the CNN models.
引用
收藏
页数:16
相关论文
共 50 条
  • [1] Tensor Factorization for Low-Rank Tensor Completion
    Zhou, Pan
    Lu, Canyi
    Lin, Zhouchen
    Zhang, Chao
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2018, 27 (03) : 1152 - 1163
  • [2] TT-TSDF: Memory-Efficient TSDF with Low-Rank Tensor Train Decomposition
    Boyko, Alexey, I
    Matrosov, Mikhail P.
    Oseledets, Ivan, V
    Tsetserukou, Dzmitry
    Ferrer, Gonzalo
    2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, : 10116 - 10121
  • [3] A low-rank tensor reconstruction and denoising method for enhancing CNN performance
    Harikumar, Rohin
    Minkoff, Susan E.
    Lou, Yifei
    2024 IEEE SOUTHWEST SYMPOSIUM ON IMAGE ANALYSIS AND INTERPRETATION, SSIAI, 2024, : 69 - 72
  • [4] A Weighted Tensor Factorization Method for Low-rank Tensor Completion
    Cheng, Miaomiao
    Jing, Liping
    Ng, Michael K.
    2019 IEEE FIFTH INTERNATIONAL CONFERENCE ON MULTIMEDIA BIG DATA (BIGMM 2019), 2019, : 30 - 38
  • [5] TEC-CNN: Toward Efficient Compressing of Convolutional Neural Nets with Low-rank Tensor Decomposition
    Wang, Yifan
    Feng, Liang
    Cai, Fenglin
    Li, Lusi
    Wu, Rui
    Li, Jie
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2025, 21 (02)
  • [6] Unifying tensor factorization and tensor nuclear norm approaches for low-rank tensor completion
    Du, Shiqiang
    Xiao, Qingjiang
    Shi, Yuqing
    Cucchiara, Rita
    Ma, Yide
    NEUROCOMPUTING, 2021, 458 : 204 - 218
  • [7] Adaptive Rank Estimation Based Tensor Factorization Algorithm for Low-Rank Tensor Completion
    Liu, Han
    Liu, Jing
    Su, Liyu
    PROCEEDINGS OF THE 38TH CHINESE CONTROL CONFERENCE (CCC), 2019, : 3444 - 3449
  • [8] PARALLEL MATRIX FACTORIZATION FOR LOW-RANK TENSOR COMPLETION
    Xu, Yangyang
    Hao, Ruru
    Yin, Wotao
    Su, Zhixun
    INVERSE PROBLEMS AND IMAGING, 2015, 9 (02) : 601 - 624
  • [9] Multimodal Core Tensor Factorization and its Applications to Low-Rank Tensor Completion
    Zeng, Haijin
    Xue, Jize
    Luong, Hiap Q.
    Philips, Wilfried
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 7010 - 7024
  • [10] CROSS: EFFICIENT LOW-RANK TENSOR COMPLETION
    Zhang, Anru
    ANNALS OF STATISTICS, 2019, 47 (02): : 936 - 964