WRA-SS: A High-Performance Accelerator Integrating Winograd With Structured Sparsity for Convolutional Neural Networks

被引:1
|
作者
Yang, Chen [1 ]
Meng, Yishuo [1 ]
Xi, Jiawei [1 ]
Xiang, Siwei [1 ]
Wang, Jianfei [1 ]
Mei, Kuizhi [1 ]
机构
[1] Xi An Jiao Tong Univ, Sch Microelect, Xian 710049, Shaanxi, Peoples R China
基金
中国国家自然科学基金;
关键词
Convolutional neural network; data storage and schedule; sparse neural network; Winograd algorithm; CNN ACCELERATOR; EFFICIENT; ALGORITHM;
D O I
10.1109/TVLSI.2023.3330993
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Sparsification for convolutional neural networks (CNNs) and convolution acceleration algorithms such as the Winograd algorithm are two efficient ways to reduce the intensive computations of existing CNNs. To better combine the sparsification and Winograd algorithm, a close integration method is proposed to dynamically reduce the invalid parameters following the Winograd transformation. To address the limitation of data bandwidth, a hierarchical two-level storage structure and corresponding data scheduling scheme are proposed, which can realize a conflict-free scheduling process. In addition, an algorithm hardware codesign method is proposed to efficiently and flexibly reduce the invalid computations led by the previous filter decomposition method. The accelerator is evaluated on Xilinx XCVU9P FPGA, reaching 412-MHz clock frequency. Compared to state-of-the-art designs, WRA-SS can achieve 1.54-5.33x and 1.17-7.39x performance improvement for VGG-16 under 80% weight sparsity and 0% weight sparsity, respectively.
引用
收藏
页码:164 / 177
页数:14
相关论文
共 50 条
  • [1] High-Performance Winograd Based Accelerator Architecture for Convolutional Neural Network
    Vardhana, M.
    Pinto, Rohan
    IEEE COMPUTER ARCHITECTURE LETTERS, 2025, 24 (01) : 21 - 24
  • [2] An Efficient Accelerator with Winograd for Novel Convolutional Neural Networks
    Lin, Zhijian
    Zhang, Meng
    Weng, Dongpeng
    Liu, Fei
    2022 5TH INTERNATIONAL CONFERENCE ON CIRCUITS, SYSTEMS AND SIMULATION (ICCSS 2022), 2022, : 126 - 130
  • [3] A High-Performance Accelerator for Large-Scale Convolutional Neural Networks
    Sun, Fan
    Wang, Chao
    Gong, Lei
    Xu, Chongchong
    Zhang, Yiwei
    Lu, Yuntao
    Li, Xi
    Zhou, Xuehai
    2017 15TH IEEE INTERNATIONAL SYMPOSIUM ON PARALLEL AND DISTRIBUTED PROCESSING WITH APPLICATIONS AND 2017 16TH IEEE INTERNATIONAL CONFERENCE ON UBIQUITOUS COMPUTING AND COMMUNICATIONS (ISPA/IUCC 2017), 2017, : 622 - 629
  • [4] A Winograd-Based Integrated Photonics Accelerator for Convolutional Neural Networks
    Mehrabian, Armin
    Miscuglio, Mario
    Alkabani, Yousra
    Sorger, Volker J.
    El-Ghazawi, Tarek
    IEEE JOURNAL OF SELECTED TOPICS IN QUANTUM ELECTRONICS, 2020, 26 (01)
  • [5] SpWA: An Efficient Sparse Winograd Convolutional Neural Networks Accelerator on FPGAs
    Lu, Liqiang
    Liang, Yun
    2018 55TH ACM/ESDA/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2018,
  • [6] High Speed Winograd Convolutional Circuit for Convolutional Neural Networks
    Chen, Ting-Wei
    Hsieh, Hao-An
    Fan, Yu-Cheng
    2022 IEEE INTERNATIONAL CONFERENCE ON CONSUMER ELECTRONICS - TAIWAN, IEEE ICCE-TW 2022, 2022, : 347 - 348
  • [7] WRA: A 2.2-to-6.3 TOPS Highly Unified Dynamically Reconfigurable Accelerator Using a Novel Winograd Decomposition Algorithm for Convolutional Neural Networks
    Yang, Chen
    Wang, Yizhou
    Wang, Xiaoli
    Geng, Li
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2019, 66 (09) : 3480 - 3493
  • [8] A LOW-LATENCY SPARSE-WINOGRAD ACCELERATOR FOR CONVOLUTIONAL NEURAL NETWORKS
    Wang, Haonan
    Liu, Wenjian
    Xu, Tianyi
    Lin, Jun
    Wang, Zhongfeng
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 1448 - 1452
  • [9] A High-efficiency FPGA-based Accelerator for Convolutional Neural Networks using Winograd Algorithm
    Huang, Y.
    Shen, J.
    Wang, Z.
    Wen, M.
    Zhang, C.
    2018 INTERNATIONAL CONFERENCE ON ELECTRONICS, COMMUNICATIONS AND CONTROL ENGINEERING (ICECC), 2018, 1026
  • [10] High performance reconfigurable accelerator for deep convolutional neural networks
    Qiao R.
    Chen G.
    Gong G.
    Lu H.
    Xi'an Dianzi Keji Daxue Xuebao/Journal of Xidian University, 2019, 46 (03): : 130 - 139