SASL: Saliency-Adaptive Sparsity Learning for Neural Network Acceleration

被引:21
|
作者
Shi, Jun [1 ]
Xu, Jianfeng [2 ]
Tasaka, Kazuyuki [2 ]
Chen, Zhibo [1 ]
机构
[1] Univ Sci & Technol China, CAS Key Lab Technol Geospatial Informat Proc & Ap, Hefei 230027, Peoples R China
[2] KDDI Res Inc, Fujimino 3568502, Japan
关键词
Training; Acceleration; Biological neural networks; Optimization; Computational modeling; Predictive models; Convolutional neural network (CNN); sparsity learning; adaptive; acceleration; compression;
D O I
10.1109/TCSVT.2020.3013170
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Accelerating the inference of CNNs is critical to their deployment in real-world applications. Among all pruning approaches, the methods of implementing a sparsity learning framework have shown effectiveness as they learn and prune the models in an end-to-end data-driven manner. However, these works impose the same sparsity regularization on all filters indiscriminately, which can hardly result in an optimal structure-sparse network. In this paper, we propose a Saliency-Adaptive Sparsity Learning (SASL) approach for further optimization. A novel and effective estimation of each filter, i.e., saliency, is designed, which is measured from two aspects: the importance for prediction performance and the consumed computational resources. During sparsity learning, the regularization strength is adjusted according to the saliency, so our optimized format can better preserve the prediction performance while zeroing out more computation-heavy filters. The calculation for saliency introduces minimum overhead to the training process, which means our SASL is very efficient. During the pruning phase, in order to optimize the proposed data-dependent criterion, a hard sample mining strategy is utilized, which shows higher effectiveness and efficiency. Extensive experiments demonstrate the superior performance of our method. Notably, on ILSVRC-2012 dataset, our approach can reduce 49.7% FLOPs of ResNet-50 with very negligible 0.39% top-1 and 0.05% top-5 accuracy degradation.
引用
收藏
页码:2008 / 2019
页数:12
相关论文
共 50 条
  • [1] Adaptive Neural Connections for Sparsity Learning
    Gain, Alex
    Kaushik, Prakhar
    Siegelmann, Hava
    2020 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2020, : 3177 - 3182
  • [2] Deep Neural Network Acceleration Method Based on Sparsity
    He, Ming
    Zhao, Haiwu
    Wang, Guozhong
    Chen, Yu
    Zhu, Linlin
    Gao, Yuan
    DIGITAL TV AND MULTIMEDIA COMMUNICATION, 2019, 1009 : 133 - 145
  • [3] ACSL: Adaptive correlation-driven sparsity learning for deep neural network compression
    He, Wei
    Wu, Meiqing
    Lam, Siew-Kei
    Neural Networks, 2021, 144 : 465 - 477
  • [4] ACSL: Adaptive correlation-driven sparsity learning for deep neural network compression
    He, Wei
    Wu, Meiqing
    Lam, Siew-Kei
    NEURAL NETWORKS, 2021, 144 : 465 - 477
  • [5] Adaptive Runtime Exploiting Sparsity in Tensor of Deep Learning Neural Network on Heterogeneous Systems
    Peng, Kuo-You
    Fu, Sheng-Yu
    Liu, Yu-Ping
    Hsu, Wei-Chung
    INTERNATIONAL CONFERENCE ON EMBEDDED COMPUTER SYSTEMS: ARCHITECTURES, MODELING, AND SIMULATION (SAMOS 2017), 2017, : 105 - 112
  • [6] Acceleration of Neural Network Learning by GPGPU
    Tsuchida, Yuta
    Yoshioka, Michifumi
    ELECTRONICS AND COMMUNICATIONS IN JAPAN, 2013, 96 (08) : 59 - 66
  • [7] Backpropagation With Sparsity Regularization for Spiking Neural Network Learning
    Yan, Yulong
    Chu, Haoming
    Jin, Yi
    Huan, Yuxiang
    Zou, Zhuo
    Zheng, Lirong
    FRONTIERS IN NEUROSCIENCE, 2022, 16
  • [8] On Acceleration of Incremental Learning in Chaotic Neural Network
    Deguchi, Toshinori
    Takahashi, Toshiki
    Ishii, Naohiro
    ADVANCES IN COMPUTATIONAL INTELLIGENCE, PT II, 2015, 9095 : 370 - 379
  • [9] Learning Hierarchical Image Representation with Sparsity, Saliency and Locality
    Yang, Jimei
    Yang, Ming-Hsuan
    PROCEEDINGS OF THE BRITISH MACHINE VISION CONFERENCE 2011, 2011,
  • [10] SASCHA-Sparsity-Aware Stochastic Computing Hardware Architecture for Neural Network Acceleration
    Romaszkan, Wojciech
    Li, Tianmu
    Gupta, Puneet
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2022, 41 (11) : 4169 - 4180