SASL: Saliency-Adaptive Sparsity Learning for Neural Network Acceleration

被引:21
|
作者
Shi, Jun [1 ]
Xu, Jianfeng [2 ]
Tasaka, Kazuyuki [2 ]
Chen, Zhibo [1 ]
机构
[1] Univ Sci & Technol China, CAS Key Lab Technol Geospatial Informat Proc & Ap, Hefei 230027, Peoples R China
[2] KDDI Res Inc, Fujimino 3568502, Japan
关键词
Training; Acceleration; Biological neural networks; Optimization; Computational modeling; Predictive models; Convolutional neural network (CNN); sparsity learning; adaptive; acceleration; compression;
D O I
10.1109/TCSVT.2020.3013170
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Accelerating the inference of CNNs is critical to their deployment in real-world applications. Among all pruning approaches, the methods of implementing a sparsity learning framework have shown effectiveness as they learn and prune the models in an end-to-end data-driven manner. However, these works impose the same sparsity regularization on all filters indiscriminately, which can hardly result in an optimal structure-sparse network. In this paper, we propose a Saliency-Adaptive Sparsity Learning (SASL) approach for further optimization. A novel and effective estimation of each filter, i.e., saliency, is designed, which is measured from two aspects: the importance for prediction performance and the consumed computational resources. During sparsity learning, the regularization strength is adjusted according to the saliency, so our optimized format can better preserve the prediction performance while zeroing out more computation-heavy filters. The calculation for saliency introduces minimum overhead to the training process, which means our SASL is very efficient. During the pruning phase, in order to optimize the proposed data-dependent criterion, a hard sample mining strategy is utilized, which shows higher effectiveness and efficiency. Extensive experiments demonstrate the superior performance of our method. Notably, on ILSVRC-2012 dataset, our approach can reduce 49.7% FLOPs of ResNet-50 with very negligible 0.39% top-1 and 0.05% top-5 accuracy degradation.
引用
收藏
页码:2008 / 2019
页数:12
相关论文
共 50 条
  • [31] A Sparsity Promoting Adaptive Algorithm for Distributed Learning
    Chouvardas, Symeon
    Slavakis, Konstantinos
    Kopsinis, Yannis
    Theodoridis, Sergios
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2012, 60 (10) : 5412 - 5425
  • [32] Deep neural network model of haptic saliency
    Anna Metzger
    Matteo Toscani
    Arash Akbarinia
    Matteo Valsecchi
    Knut Drewing
    Scientific Reports, 11
  • [33] Deep neural network model of haptic saliency
    Metzger, Anna
    Toscani, Matteo
    Akbarinia, Arash
    Valsecchi, Matteo
    Drewing, Knut
    SCIENTIFIC REPORTS, 2021, 11 (01)
  • [34] Convolutional Neural Network for Saliency Detection in Images
    Misaghi, Hooman
    Moghadam, Reza Askari
    Madani, Kurosh
    2018 6TH IRANIAN JOINT CONGRESS ON FUZZY AND INTELLIGENT SYSTEMS (CFIS), 2018, : 17 - 19
  • [35] A neural network implementation of a saliency map model
    de Brecht, Matthew
    Saiki, Jun
    NEURAL NETWORKS, 2006, 19 (10) : 1467 - 1474
  • [36] Learning Structured Sparsity in Deep Neural Networks
    Wen, Wei
    Wu, Chunpeng
    Wang, Yandan
    Chen, Yiran
    Li, Hai
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29
  • [37] Design of a Neural Network Acceleration Autopilot for Spinning Projectile Based on Adaptive Disturbance Observer
    Wang, Wei
    Yang, Jing
    Nan, Yuxiang
    Li, Junhui
    Wang, Yuchen
    Binggong Xuebao/Acta Armamentarii, 2024, 45 (11): : 3841 - 3855
  • [38] Cache-locality Based Adaptive Warp Scheduling for Neural Network Acceleration on GPGPUs
    Hu, Weiming
    Zhou, Yi
    Quan, Ying
    Wang, Yuanfeng
    Lou, Xin
    2022 IEEE 35TH INTERNATIONAL SYSTEM-ON-CHIP CONFERENCE (IEEE SOCC 2022), 2022, : 190 - 195
  • [39] Adaptive Convolutional Neural Network for Predicting Steering Angle and Acceleration on Autonomous Driving Scenario
    Vasiljevic, Ive
    Music, Josip
    Mendes, Joa
    Lima, Jose
    OPTIMIZATION, LEARNING ALGORITHMS AND APPLICATIONS, PT II, OL2A 2023, 2024, 1982 : 132 - 147
  • [40] Explainable Online Deep Neural Network Selection Using Adaptive Saliency Maps for Time Series Forecasting
    Saadallah, Amal
    Jakobs, Matthias
    Morik, Katharina
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, 2021, 12975 : 404 - 420