Systematic Literature Review on Cost-Efficient Deep Learning

被引:5
作者
Klemetti, Antti [1 ]
Raatikainen, Mikko [1 ]
Myllyaho, Lalli [1 ]
Mikkonen, Tommi [2 ]
Nurminen, Jukka K. [1 ]
机构
[1] Univ Helsinki, Fac Sci, Dept Comp Sci, Helsinki 00014, Finland
[2] Univ Jyvaskyla, Fac Informat Technol, Jyvaskyla 40014, Finland
关键词
Costs; Cloud computing; Deep learning; Training; Neurons; Computational modeling; Business; cost-efficiency; cost reduction; deep learning; deep neural network; edge offloading; machine learning; systematic literature review; NEURAL-NETWORKS; HARDWARE;
D O I
10.1109/ACCESS.2023.3275431
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Cloud computing and deep learning, the recent trends in the software industry, have enabled small companies to scale their business up rapidly. However, this growth is not without a cost - deep learning models are related to the heaviest workloads in cloud data centers. When the business grows, the monetary cost of deep learning in the cloud also grows fast. Deep learning practitioners should be prepared and equipped to limit the growing cost. We emphasize monetary cost instead of computational cost although often the same methods decrease both types of cost. We performed a systematic literature review on the methods to control the cost of deep learning. Our library search resulted in 16,066 papers from three article databases, IEEE Xplore, ACM Digital Library, and Scopus. We narrowed them down to 112 papers that we categorized and summarized. We found that: 1) Optimizing inference has raised more interest than optimizing training. Widely used deep learning libraries already support inference optimization methods, such as quantization, pruning, and teacher-student. 2) The research has been centered around image inputs, and there seems to be a research gap for other types of inputs. 3) The research has been hardware-oriented, and the most typical approach to control the cost of deep learning is based on algorithm-hardware co-design. 4) Offloading some of the processing to client devices is gaining interest and can potentially reduce the monetary cost of deep learning.
引用
收藏
页码:90158 / 90180
页数:23
相关论文
共 55 条
  • [41] Kipf TN, 2017, Arxiv, DOI arXiv:1609.02907
  • [42] Ruder S, 2017, Arxiv, DOI [arXiv:1609.04747, DOI 10.48550/ARXIV.1609.04747, 10.48550/ARXIV.1609.04747]
  • [43] The Case for VM-Based Cloudlets in Mobile Computing
    Satyanarayanan, Mahadev
    Bahl, Paramvir
    Caceres, Ramon
    Davies, Nigel
    [J]. IEEE PERVASIVE COMPUTING, 2009, 8 (04) : 14 - 23
  • [44] Green AI
    Schwartz, Roy
    Dodge, Jesse
    Smith, Noah A.
    Etzioni, Oren
    [J]. COMMUNICATIONS OF THE ACM, 2020, 63 (12) : 54 - 63
  • [45] Shafique M, 2018, DES AUT TEST EUROPE, P827, DOI 10.23919/DATE.2018.8342120
  • [46] FPGA-Based Accelerators of Deep Learning Networks for Learning and Classification: A Review
    Shawahna, Ahmad
    Sait, Sadiq M.
    El-Maleh, Aiman
    [J]. IEEE ACCESS, 2019, 7 : 7823 - 7859
  • [47] Sung WY, 2016, Arxiv, DOI arXiv:1511.06488
  • [48] Sze V, 2018, IEEE CUST INTEGR CIR
  • [49] Efficient Processing of Deep Neural Networks: A Tutorial and Survey
    Sze, Vivienne
    Chen, Yu-Hsin
    Yang, Tien-Ju
    Emer, Joel S.
    [J]. PROCEEDINGS OF THE IEEE, 2017, 105 (12) : 2295 - 2329
  • [50] TensorFlow Lite, 2022, ML for Mobile and Edge Devices