Towards Efficient Model Compression via Learned Global Ranking

被引:108
作者
Chin, Ting-Wu [1 ]
Ding, Ruizhou [1 ]
Zhang, Cha [2 ]
Marculescu, Diana [1 ,3 ]
机构
[1] Carnegie Mellon Univ, Pittsburgh, PA 15213 USA
[2] Microsoft Cloud & AI, Redmond, WA USA
[3] Univ Texas Austin, Austin, TX 78712 USA
来源
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR) | 2020年
基金
美国国家科学基金会;
关键词
D O I
10.1109/CVPR42600.2020.00159
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Pruning convolutional filters has demonstrated its effectiveness in compressing ConvNets. Prior art in filter pruning requires users to specify a target model complexity (e.g., model size or FLOP count) for the resulting architecture. However, determining a target model complexity can be difficult for optimizing various embodied AI applications such as autonomous robots, drones, and user-facing applications. First, both the accuracy and the speed of ConvNets can affect the performance of the application. Second, the performance of the application can be hard to assess without evaluating ConvNets during inference. As a consequence, finding a sweet-spot between the accuracy and speed via filter pruning, which needs to be done in a trialand-error fashion, can be time-consuming. This work takes a first step toward making this process more efficient by altering the goal of model compression to producing a set of ConvNets with various accuracy and latency trade-offs instead of producing one ConvNet targeting some pre-defined latency constraint. To this end, we propose to learn a global ranking of the filters across different layers of the ConvNet, which is used to obtain a set of ConvNet architectures that have different accuracy/latency trade-offs by pruning the bottom-ranked filters. Our proposed algorithm, LeGR, is shown to be 2 x to 3 x faster than prior work while having comparable or better performance when targeting seven pruned ResNet-56 with different accuracy/FLOPs profiles on the CIFAR-100 dataset. Additionally, we have evaluated LeGR on ImageNet and Bird-200 with ResNet-50 and MobileNetV2 to demonstrate its effectiveness. Code available at https://github.com/cmu-enyac/LeGR.
引用
收藏
页码:1515 / 1525
页数:11
相关论文
共 72 条
  • [1] Adam H, 2018, ARXIV180403230
  • [2] Vision-and-Language Navigation: Interpreting visually-grounded navigation instructions in real environments
    Anderson, Peter
    Wu, Qi
    Teney, Damien
    Bruce, Jake
    Johnson, Mark
    Sunderhauf, Niko
    Reid, Ian
    Gould, Stephen
    van den Hengel, Anton
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 3674 - 3683
  • [3] [Anonymous], 2018, ARXIV180608198
  • [4] [Anonymous], 2018, ARXIV180711626
  • [5] [Anonymous], ARXIV190402877
  • [6] [Anonymous], 2017, ARXIV170404861
  • [7] [Anonymous], 2017, P 31 C NEUR INF PROC
  • [8] [Anonymous], 2019, ARXIV190401201
  • [9] [Anonymous], 2018, INT C LEARN REPR
  • [10] [Anonymous], 2019, ARXIV190202910