Efficient Use of GPU Memory for Large-Scale Deep Learning Model Training

被引:13
|
作者
Choi, Hyeonseong [1 ]
Lee, Jaehwan [1 ]
机构
[1] Korea Aerosp Univ, Sch Elect & Informat Engn, Goyang Si 10540, South Korea
来源
APPLIED SCIENCES-BASEL | 2021年 / 11卷 / 21期
基金
新加坡国家研究基金会;
关键词
deep learning; large-scale model; CUDA Unified Memory; PyTorch;
D O I
10.3390/app112110377
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
To achieve high accuracy when performing deep learning, it is necessary to use a large-scale training model. However, due to the limitations of GPU memory, it is difficult to train large-scale training models within a single GPU. NVIDIA introduced a technology called CUDA Unified Memory with CUDA 6 to overcome the limitations of GPU memory by virtually combining GPU memory and CPU memory. In addition, in CUDA 8, memory advise options are introduced to efficiently utilize CUDA Unified Memory. In this work, we propose a newly optimized scheme based on CUDA Unified Memory to efficiently use GPU memory by applying different memory advise to each data type according to access patterns in deep learning training. We apply CUDA Unified Memory technology to PyTorch to see the performance of large-scale learning models through the expanded GPU memory. We conduct comprehensive experiments on how to efficiently utilize Unified Memory by applying memory advises when performing deep learning. As a result, when the data used for deep learning are divided into three types and a memory advise is applied to the data according to the access pattern, the deep learning execution time is reduced by 9.4% compared to the default Unified Memory.
引用
收藏
页数:17
相关论文
共 50 条
  • [31] Large-Scale Mobile App Identification Using Deep Learning
    Rezaei, Shahbaz
    Kroencke, Bryce
    Liu, Xin
    IEEE ACCESS, 2020, 8 : 348 - 362
  • [32] Rich Punctuations Prediction Using Large-scale Deep Learning
    Wu, Xueyang
    Zhu, Su
    Wu, Yue
    Yu, Kai
    2016 10TH INTERNATIONAL SYMPOSIUM ON CHINESE SPOKEN LANGUAGE PROCESSING (ISCSLP), 2016,
  • [33] Deep Learning Hyperspectral Pansharpening on Large-Scale PRISMA Dataset
    Zini, Simone
    Barbato, Mirko Paolo
    Piccoli, Flavio
    Napoletano, Paolo
    REMOTE SENSING, 2024, 16 (12)
  • [34] Hybrid Beamforming With Deep Learning for Large-Scale Antenna Arrays
    Hu, Rentao
    Jiang, Lijun
    Li, Ping
    IEEE ACCESS, 2021, 9 : 54690 - 54699
  • [35] An Efficient and Non-Intrusive GPU Scheduling Framework for Deep Learning Training Systems
    Wang, Shaoqi
    Gonzalez, Oscar J.
    Zhou, Xiaobo
    Williams, Thomas
    Friedman, Brian D.
    Havemann, Martin
    Woo, Thomas
    PROCEEDINGS OF SC20: THE INTERNATIONAL CONFERENCE FOR HIGH PERFORMANCE COMPUTING, NETWORKING, STORAGE AND ANALYSIS (SC20), 2020,
  • [36] A large-scale evaluation framework for EEG deep learning architectures
    Heilmeyer, Felix A.
    Schirrmeister, Robin T.
    Fiederer, Lukas D. J.
    Voelker, Martin
    Behncke, Joos
    Ball, Tonio
    2018 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2018, : 1039 - 1045
  • [37] Analysis of large deviations behavior of multi-GPU memory access in deep learning
    P. S. Tamizharasan
    N. Ramasubramanian
    The Journal of Supercomputing, 2018, 74 : 2199 - 2212
  • [38] MixPipe: Efficient Bidirectional Pipeline Parallelism for Training Large-Scale Models
    Zhang, Weigang
    Zhou, Biyu
    Tang, Xuehai
    Wang, Zhaoxing
    Hu, Songlin
    2023 60TH ACM/IEEE DESIGN AUTOMATION CONFERENCE, DAC, 2023,
  • [39] Deep learning model with low-dimensional random projection for large-scale image search
    Alzu'bi, Ahmad
    Abuarqoub, Abdelrahman
    ENGINEERING SCIENCE AND TECHNOLOGY-AN INTERNATIONAL JOURNAL-JESTECH, 2020, 23 (04): : 911 - 920
  • [40] Learning to Train and to Explain a Deep Survival Model with Large-Scale Ovarian Cancer Transcriptomic Data
    Menand, Elena Spirina
    De Vries-Brilland, Manon
    Tessier, Leslie
    Dauve, Jonathan
    Campone, Mario
    Verriele, Veronique
    Jrad, Nisrine
    Marion, Jean-Marie
    Chauvet, Pierre
    Passot, Christophe
    Morel, Alain
    BIOMEDICINES, 2024, 12 (12)