Global Channel Pruning With Self-Supervised Mask Learning

被引:1
|
作者
Ma, Ming [1 ]
Zhang, Tongzhou [1 ]
Wang, Ziming [1 ]
Wang, Yue [1 ]
Du, Taoli [1 ]
Li, Wenhui [1 ]
机构
[1] Jilin Univ, Coll Comp Sci & Technol, Changchun 130012, Peoples R China
关键词
Self-supervised learning; Training; Filters; Sparse matrices; Supervised learning; Neural networks; Circuits and systems; Accuracy; Time series analysis; Libraries; Deep neural networks; network pruning; self-supervised learning;
D O I
10.1109/TCSVT.2024.3488098
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Network pruning is widely used in model compression due to its simplicity and efficiency. Existing methods typically introduce sparse loss regularization to learn masks. However, this sparse regularization approach lacks a clear criterion for evaluating channel importance and relies on manually defined rules, leading to a decline in model performance. In this article, a Self-Supervised Mask Learning (SSML) method for global channel pruning is proposed, casting mask learning as a self-supervised binary classification task to automatically identify less important channels. Specifically, a dedicated pretext task is designed for the channelwise masks, which leverages the original network to generate pseudo-labels from the mask itself to guide mask learning. Then, a polarization mask loss function is proposed, transforming the discrete mask learning problem into a differentiable binary classification problem. The proposed loss function distinguishes the similarity between pseudo-labels and masks, clustering similar masks together in the feature space and separating dissimilar masks, ultimately allowing channels with masks of 0 to be safely removed without damaging the performance of the pruned model. In addition, SSML can train from scratch to yield a compact model. Extensive experiments on CIFAR-10, CIFAR-100 and ImageNet datasets demonstrate that SSML outperforms state-of-the-art methods. For instance, SSML prunes 52.7% of the FLOPs of ResNe34 on the ImageNet dataset with only 0.01% drop in Top-1 accuracy. Moreover, the generalization of SSML is verified on downstream tasks.
引用
收藏
页码:2013 / 2025
页数:13
相关论文
共 50 条
  • [1] Fire TogetherWire Together: A Dynamic Pruning Approach with Self-Supervised Mask Prediction
    Elkerdawy, Sara
    Elhoushi, Mostafa
    Zhang, Hong
    Ray, Nilanjan
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 12444 - 12453
  • [2] Research on Mask RCNN based on rotating self-supervised learning
    Wang, Xuedong
    Su, Shi
    Huang, Hongcheng
    Chu, Pengzhi
    2022 41ST CHINESE CONTROL CONFERENCE (CCC), 2022, : 6978 - 6981
  • [3] Self-Supervised Visual Representations Learning by Contrastive Mask Prediction
    Zhao, Yucheng
    Wang, Guangting
    Luo, Chong
    Zeng, Wenjun
    Zha, Zheng-Jun
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 10140 - 10149
  • [4] Unified Mask Embedding and Correspondence Learning for Self-Supervised Video Segmentation
    Li, Liulei
    Wang, Wenguan
    Zhou, Tianfei
    Li, Jianwu
    Yang, Yi
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 18706 - 18716
  • [5] Embedding Global Contrastive and Local Location in Self-Supervised Learning
    Zhao, Wenyi
    Li, Chongyi
    Zhang, Weidong
    Yang, Lu
    Zhuang, Peixian
    Li, Lingqiao
    Fan, Kefeng
    Yang, Huihua
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (05) : 2275 - 2289
  • [6] PointCMP: Contrastive Mask Prediction for Self-supervised Learning on Point Cloud Videos
    Shen, Zhiqiang
    Sheng, Xiaoxiao
    Wang, Longguang
    Guo, Yulan
    Liu, Qiong
    Zhou, Xi
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 1212 - 1222
  • [7] Self-Supervised Point Cloud Understanding via Mask Transformer and Contrastive Learning
    Wang, Di
    Yang, Zhi-Xin
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2023, 8 (01) : 184 - 191
  • [8] Gated Self-supervised Learning for Improving Supervised Learning
    Fuadi, Erland Hillman
    Ruslim, Aristo Renaldo
    Wardhana, Putu Wahyu Kusuma
    Yudistira, Novanto
    2024 IEEE CONFERENCE ON ARTIFICIAL INTELLIGENCE, CAI 2024, 2024, : 611 - 615
  • [9] Self-Supervised Dialogue Learning
    Wu, Jiawei
    Wang, Xin
    Wang, William Yang
    57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 3857 - 3867
  • [10] Longitudinal self-supervised learning
    Zhao, Qingyu
    Liu, Zixuan
    Adeli, Ehsan
    Pohl, Kilian M.
    MEDICAL IMAGE ANALYSIS, 2021, 71