SASL: Saliency-Adaptive Sparsity Learning for Neural Network Acceleration

被引:21
|
作者
Shi, Jun [1 ]
Xu, Jianfeng [2 ]
Tasaka, Kazuyuki [2 ]
Chen, Zhibo [1 ]
机构
[1] Univ Sci & Technol China, CAS Key Lab Technol Geospatial Informat Proc & Ap, Hefei 230027, Peoples R China
[2] KDDI Res Inc, Fujimino 3568502, Japan
关键词
Training; Acceleration; Biological neural networks; Optimization; Computational modeling; Predictive models; Convolutional neural network (CNN); sparsity learning; adaptive; acceleration; compression;
D O I
10.1109/TCSVT.2020.3013170
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Accelerating the inference of CNNs is critical to their deployment in real-world applications. Among all pruning approaches, the methods of implementing a sparsity learning framework have shown effectiveness as they learn and prune the models in an end-to-end data-driven manner. However, these works impose the same sparsity regularization on all filters indiscriminately, which can hardly result in an optimal structure-sparse network. In this paper, we propose a Saliency-Adaptive Sparsity Learning (SASL) approach for further optimization. A novel and effective estimation of each filter, i.e., saliency, is designed, which is measured from two aspects: the importance for prediction performance and the consumed computational resources. During sparsity learning, the regularization strength is adjusted according to the saliency, so our optimized format can better preserve the prediction performance while zeroing out more computation-heavy filters. The calculation for saliency introduces minimum overhead to the training process, which means our SASL is very efficient. During the pruning phase, in order to optimize the proposed data-dependent criterion, a hard sample mining strategy is utilized, which shows higher effectiveness and efficiency. Extensive experiments demonstrate the superior performance of our method. Notably, on ILSVRC-2012 dataset, our approach can reduce 49.7% FLOPs of ResNet-50 with very negligible 0.39% top-1 and 0.05% top-5 accuracy degradation.
引用
收藏
页码:2008 / 2019
页数:12
相关论文
共 50 条
  • [41] DEEP NEURAL NETWORK BASED ADAPTIVE LEARNING FOR SWITCHED SYSTEMS
    He, Junjie
    Xu, Zhihang
    Liao, Qifeng
    DISCRETE AND CONTINUOUS DYNAMICAL SYSTEMS-SERIES S, 2023, 16 (07): : 1827 - 1855
  • [42] Neural network indirect adaptive control with fast learning algorithm
    Jeon, GJ
    Lee, I
    NEUROCOMPUTING, 1996, 13 (2-4) : 185 - 199
  • [43] A neural network approach for learning image similarity in adaptive CBIR
    Muneesawang, P
    Guan, L
    2001 IEEE FOURTH WORKSHOP ON MULTIMEDIA SIGNAL PROCESSING, 2001, : 257 - 262
  • [44] Adaptive sliding mode approach for learning in a feedforward neural network
    Yu, X
    Zhihong, M
    Rahman, SMM
    NEURAL COMPUTING & APPLICATIONS, 1998, 7 (04): : 289 - 294
  • [45] Adaptive sliding mode approach for learning in a feedforward neural network
    X. Yu
    M. Zhihong
    S. M. Monzurur Rahman
    Neural Computing & Applications, 1998, 7 : 289 - 294
  • [46] Integration of Neural Network Algorithm in Adaptive Learning Management System
    Lagman, Ace C.
    Alcober, Geliza Marie I.
    Fernando, Ma Corazon G.
    Goh, Marie Luvett I.
    Lalata, Jay-ar P.
    Ortega, John Heland Jasper C.
    Perez, Maria Rona L.
    Solomo, Maria Vicky S.
    Claour, Julius P.
    PROCEEDINGS OF 2020 3RD INTERNATIONAL CONFERENCE ON ROBOT SYSTEMS AND APPLICATIONS, ICRSA2020, 2020, : 82 - 87
  • [47] Hierarchical Fusion Evolving Spiking Neural Network for Adaptive Learning
    Al Zoubi, Obada
    Mayeli, Ahmad
    Awad, Mariette
    Retai, Hazem
    PROCEEDINGS OF 2018 IEEE 17TH INTERNATIONAL CONFERENCE ON COGNITIVE INFORMATICS & COGNITIVE COMPUTING (ICCI*CC 2018), 2018, : 86 - 91
  • [48] Dynamic sparsity and model feature learning enhanced training for convolutional neural network-pruning
    Ruan X.
    Hu W.
    Liu Y.
    Li B.
    Zhongguo Kexue Jishu Kexue/Scientia Sinica Technologica, 2022, 52 (05): : 667 - 681
  • [49] DEEP NEURAL NETWORK BASED ADAPTIVE LEARNING FOR SWITCHED SYSTEMS
    He, Junjie
    Xu, Zhihang
    Liao, Qifeng
    DISCRETE AND CONTINUOUS DYNAMICAL SYSTEMS-SERIES S, 2023,
  • [50] Adaptive learning cost-sensitive convolutional neural network
    Hou, Yun
    Fan, Hong
    Li, Li
    Li, Bailin
    IET COMPUTER VISION, 2021, 15 (05) : 346 - 355