Performance-Aware Approximation of Global Channel Pruning for Multitask CNNs

被引:26
作者
Ye, Hancheng [1 ]
Zhang, Bo [3 ]
Chen, Tao [1 ]
Fan, Jiayuan [2 ]
Wang, Bin [1 ]
机构
[1] Fudan Univ, Sch Informat Sci & Technol, Shanghai 200433, Peoples R China
[2] Fudan Univ, Acad Engn & Technol, Shanghai 200433, Peoples R China
[3] Shanghai AI Lab, Shanghai 200232, Peoples R China
基金
中国国家自然科学基金;
关键词
Channel pruning; multitask learning; performance-aware oracle criterion; sequentially greedy algorithm; DEEP; MODELS;
D O I
10.1109/TPAMI.2023.3260903
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Global channel pruning (GCP) aims to remove a subset of channels (filters) across different layers from a deep model without hurting the performance. Previous works focus on either single task model pruning or simply adapting it to multitask scenario, and still face the following problems when handling multitask pruning: 1) Due to the task mismatch, a well-pruned backbone for classification task focuses on preserving filters that can extract category-sensitive information, causing filters that may be useful for other tasks to be pruned during the backbone pruning stage; 2) For multitask predictions, different filters within or between layers are more closely related and interacted than that for single task prediction, making multitask pruning more difficult. Therefore, aiming at multitask model compression, we propose a Performance-Aware Global Channel Pruning (PAGCP) framework. We first theoretically present the objective for achieving superior GCP, by considering the joint saliency of filters from intra- and inter-layers. Then a sequentially greedy pruning strategy is proposed to optimize the objective, where a performance-aware oracle criterion is developed to evaluate sensitivity of filters to each task and preserve the globally most task-related filters. Experiments on several multitask datasets show that the proposed PAGCP can reduce the FLOPs and parameters by over 60% with minor performance drop, and achieves 1.2x similar to 3.3x acceleration on both cloud and mobile platforms. Our code is available at http://www.github.com/HankYe/PAGCP.git.
引用
收藏
页码:10267 / 10284
页数:18
相关论文
共 69 条
[1]  
Anisimov D, 2017, 2017 14TH IEEE INTERNATIONAL CONFERENCE ON ADVANCED VIDEO AND SIGNAL BASED SURVEILLANCE (AVSS)
[2]   Detect What You Can: Detecting and Representing Objects using Holistic Models and Body Parts [J].
Chen, Xianjie ;
Mottaghi, Roozbeh ;
Liu, Xiaobai ;
Fidler, Sanja ;
Urtasun, Raquel ;
Yuille, Alan .
2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2014, :1979-1986
[3]  
Chu X., 2020, INT C LEARN REPR, P1
[4]  
Chung I, 2020, PR MACH LEARN RES, V119
[5]   MAXIMUM LIKELIHOOD FROM INCOMPLETE DATA VIA EM ALGORITHM [J].
DEMPSTER, AP ;
LAIRD, NM ;
RUBIN, DB .
JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-METHODOLOGICAL, 1977, 39 (01) :1-38
[6]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[7]   ResRep: Lossless CNN Pruning via Decoupling Remembering and Forgetting [J].
Ding, Xiaohan ;
Hao, Tianxiang ;
Tan, Jianchao ;
Liu, Ji ;
Han, Jungong ;
Guo, Yuchen ;
Ding, Guiguang .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :4490-4500
[8]  
Dong XY, 2019, ADV NEUR IN, V32
[9]   The PASCAL Visual Object Classes Challenge: A Retrospective [J].
Everingham, Mark ;
Eslami, S. M. Ali ;
Van Gool, Luc ;
Williams, Christopher K. I. ;
Winn, John ;
Zisserman, Andrew .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2015, 111 (01) :98-136
[10]  
France J, 2018, HIST WARFARE, V116, P1, DOI 10.1163/9789004349599_002