Automatic Channel Pruning with Hyper-parameter Search and Dynamic Masking

被引:3
作者
Li, Baopu [1 ]
Fan, Yanwen [2 ]
Pan, Zhihong [1 ]
Bian, Yuchen [3 ]
Zhang, Gang [2 ]
机构
[1] Baidu USA LLC, Sunnyvale, CA 94089 USA
[2] Baidu Inc, VIS Dept, Beijing, Peoples R China
[3] Baidu Res, Beijing, Peoples R China
来源
PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021 | 2021年
关键词
Model compression; Network pruning; Auto ML;
D O I
10.1145/3474085.3475370
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Modern deep neural network models tend to be large and computationally intensive. One typical solution to this issue is model pruning. However, most current model pruning algorithms depend on hand crafted rules or need to input the pruning ratio beforehand. To overcome this problem, we propose a learning based automatic channel pruning algorithm for deep neural network, which is inspired by recent automatic machine learning (Auto ML). A two objectives' pruning problem that aims for the weights and the remaining channels for each layer is first formulated. An alternative optimization approach is then proposed to derive the channel numbers and weights simultaneously. In the process of pruning, we utilize a searchable hyper-parameter, remaining ratio, to denote the number of channels in each convolution layer, and then a dynamic masking process is proposed to describe the corresponding channel evolution. To adjust the trade-off between accuracy of a model and the pruning ratio of floating point operations, a new loss function is further introduced. Extensive experimental results on benchmark datasets demonstrate that our scheme achieves competitive results for neural network pruning.
引用
收藏
页码:2121 / 2129
页数:9
相关论文
共 42 条
  • [1] Aghasi Alireza, 2016, ABS161105162 CORR
  • [2] [Anonymous], 2020, CVPR
  • [3] [Anonymous], 1994, J GLOBAL OPTIMIZATIO
  • [4] [Anonymous], 2016, ARXIV160205149
  • [5] Bergstra James, 2012, J MACHINE LEARNING R, V12
  • [6] Bianco Simone, 2018, ABS181000736 CORR
  • [7] Chen YH, 2016, ISSCC DIG TECH PAP I, V59, P262, DOI 10.1109/ISSCC.2016.7418007
  • [8] Ding Ruizhou, 2020, ARXIVCSCV190412368
  • [9] Ding Xiaohan, 2019, ARXIV190912778
  • [10] Franceschi L, 2017, PR MACH LEARN RES, V70