Optimizing Memory Efficiency for Deep Convolutional Neural Network Accelerators

被引:1
|
作者
Li, Xiaowei [1 ]
Li, Jiajun
Yan, Guihai [1 ]
机构
[1] Chinese Acad Sci, Inst Comp Technol, State Key Lab Comp Architecture, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
Deep Convolutional Neural Networks; Accelerator Architecture; Memory Efficiency;
D O I
10.1166/jolpe.2018.1580
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Convolutional Neural Network (CNN) accelerators have achieved nominal performance and energy efficiency speedup compared to traditional general purpose CPU- and GPU-based solutions. Although optimizations on computation have been intensively studied, the energy efficiency of such accelerators remains limited by off-chip memory accesses since their energy cost is magnitudes higher than other operations. Minimizing off-chip memory access volume, therefore, is the key to further improving energy efficiency. The prior state-of-the-art uses rigid data reuse patterns and is sub-optimal for some, or even all, of the individual convolutional layers. To overcome the problem, this paper proposed an adaptive layer partitioning and scheduling scheme, called SmartShuttle, to minimize off-chip memory accesses for CNN accelerators. Smartshuttle can adaptively switch among different data reuse schemes and the corresponding tiling factor settings to dynamically match different convolutional layers and fully-connected layers. Moreover, SmartShuttle thoroughly investigates the impact of data reusability and sparsity on the memory access volume. The experimental results show that SmartShuttle processes the convolutional layers at 434.8 multiply and accumulations (MACs)/DRAM access for VGG16 (batch size = 3), and 526.3 MACs/DRAM access for AlexNet (batch size = 4), which outperforms the state-of-the-art approach (Eyeriss) by 52.2% and 52.6%, respectively.
引用
收藏
页码:496 / 507
页数:12
相关论文
共 50 条
  • [31] Optimizing Accelerator on FPGA for Deep Convolutional Neural Networks
    Dong, Yong
    Hu, Wei
    Wang, Yonghao
    Jiao, Qiang
    Chen, Shuang
    ALGORITHMS AND ARCHITECTURES FOR PARALLEL PROCESSING, ICA3PP 2020, PT II, 2020, 12453 : 97 - 110
  • [32] On-Chip Memory Optimization of High Efficiency Accelerator for Deep Convolutional Neural Networks
    Lai, Tzu-Yi
    Chen, Kuan-Hung
    2018 INTERNATIONAL SOC DESIGN CONFERENCE (ISOCC), 2018, : 82 - 83
  • [33] Automated optimization for memory-efficient high-performance deep neural network accelerators
    Kim, HyunMi
    Lyuh, Chun-Gi
    Kwon, Youngsu
    ETRI JOURNAL, 2020, 42 (04) : 505 - 517
  • [34] On-Chip Memory Technology Design Space Explorations for Mobile Deep Neural Network Accelerators
    Li, Haitong
    Bhargava, Mudit
    Whatmough, Paul N.
    Wong, H-S Philip
    PROCEEDINGS OF THE 2019 56TH ACM/EDAC/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2019,
  • [35] Low power & mobile hardware accelerators for deep convolutional neural networks
    Scanlan, Anthony G.
    INTEGRATION-THE VLSI JOURNAL, 2019, 65 : 110 - 127
  • [36] Dynamic Precision Multiplier For Deep Neural Network Accelerators
    Ding, Chen
    Yuxiang, Huan
    Zheng, Lirong
    Zou, Zhuo
    2020 IEEE 33RD INTERNATIONAL SYSTEM-ON-CHIP CONFERENCE (SOCC), 2020, : 180 - 184
  • [37] Domain Wall Memory-Based Design of Deep Neural Network Convolutional Layers
    Chung, Jinil
    Choi, Woong
    Park, Jongsun
    Ghosh, Swaroop
    IEEE ACCESS, 2020, 8 (08): : 19783 - 19798
  • [38] Extended Bit-Plane Compression for Convolutional Neural Network Accelerators
    Cavigelli, Lukas
    Benini, Luca
    2019 IEEE INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE CIRCUITS AND SYSTEMS (AICAS 2019), 2019, : 279 - 283
  • [39] Parallel Convolutional Neural Network (CNN) Accelerators Based on Stochastic Computing
    Zhang, Yawen
    Zhang, Xinyue
    Song, Jiahao
    Wang, Yuan
    Huang, Ru
    Wang, Runsheng
    PROCEEDINGS OF THE 2019 IEEE INTERNATIONAL WORKSHOP ON SIGNAL PROCESSING SYSTEMS (SIPS 2019), 2019, : 19 - 24
  • [40] CNNWire: Boosting Convolutional Neural Network with Winograd on ReRAM based Accelerators
    Lin, Jilan
    Li, Shuangchen
    Hu, Xing
    Deng, Lei
    Xie, Yuan
    GLSVLSI '19 - PROCEEDINGS OF THE 2019 ON GREAT LAKES SYMPOSIUM ON VLSI, 2019, : 283 - 286