Multi-view knowledge distillation for efficient semantic segmentation

被引:4
作者
Wang, Chen [1 ]
Zhong, Jiang [1 ]
Dai, Qizhu [1 ]
Qi, Yafei [2 ]
Shi, Fengyuan [3 ]
Fang, Bin [1 ]
Li, Xue [4 ]
机构
[1] Chongqing Univ, Sch Comp Sci, Chongqing 400044, Peoples R China
[2] Cent South Univ, Sch Comp Sci & Engn, Changsha 410083, Peoples R China
[3] Northeastern Univ, Sch Informat Sci & Engn, Shenyang 110819, Peoples R China
[4] Univ Queensland, Sch Informat Technol & Elect Engn, Brisbane, Qld 4072, Australia
基金
中国国家自然科学基金;
关键词
Multi-view learning; Knowledge distillation; Knowledge aggregation; Semantic segmentation; ENSEMBLE;
D O I
10.1007/s11554-023-01296-6
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Current state-of-the-art semantic segmentation models achieve remarkable success in segmentation accuracy. However, the huge model size and computing cost restrict their applications on low-latency online systems or devices. Knowledge distillation has been one popular solution for compressing large-scale segmentation models, which train a small segmentation model from a large teacher model. However, one teacher model's knowledge may be insufficiently diverse to train an accurate student model. Meanwhile, the student model may inherit bias from the teacher model. This paper proposes a multi-view knowledge distillation framework called MVKD for efficient semantic segmentation. MVKD could aggregate the multi-view knowledge from multiple teacher models and transfer the multi-view knowledge to the student model. In MVKD, we introduce one multi-view co-tuning strategy to acquire uniformity among the multi-view knowledge in features from different teachers. In addition, we propose a multi-view feature distillation loss and a multi-view output distillation loss to transfer the multi-view knowledge in the features and outputs from multiple teachers to the student. We evaluate the proposed MVKD on three benchmark datasets, Cityscapes, CamVid, and Pascal VOC 2012. Experimental results demonstrate the effectiveness of the proposed MVKD in compressing semantic segmentation models.
引用
收藏
页数:11
相关论文
共 38 条
  • [31] Yukang Wang, 2020, Computer Vision - ECCV 2020. 16th European Conference. Proceedings. Lecture Notes in Computer Science (LNCS 12352), P346, DOI 10.1007/978-3-030-58571-6_21
  • [32] Deep Mutual Learning
    Zhang, Ying
    Xiang, Tao
    Hospedales, Timothy M.
    Lu, Huchuan
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 4320 - 4328
  • [33] ICNet for Real-Time Semantic Segmentation on High-Resolution Images
    Zhao, Hengshuang
    Qi, Xiaojuan
    Shen, Xiaoyong
    Shi, Jianping
    Jia, Jiaya
    [J]. COMPUTER VISION - ECCV 2018, PT III, 2018, 11207 : 418 - 434
  • [34] Pyramid Scene Parsing Network
    Zhao, Hengshuang
    Shi, Jianping
    Qi, Xiaojuan
    Wang, Xiaogang
    Jia, Jiaya
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 6230 - 6239
  • [35] Conditional Random Fields as Recurrent Neural Networks
    Zheng, Shuai
    Jayasumana, Sadeep
    Romera-Paredes, Bernardino
    Vineet, Vibhav
    Su, Zhizhong
    Du, Dalong
    Huang, Chang
    Torr, Philip H. S.
    [J]. 2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, : 1529 - 1537
  • [36] Rethinking Semantic Segmentation from a Sequence-to-Sequence Perspective with Transformers
    Zheng, Sixiao
    Lu, Jiachen
    Zhao, Hengshuang
    Zhu, Xiatian
    Luo, Zekun
    Wang, Yabiao
    Fu, Yanwei
    Feng, Jianfeng
    Xiang, Tao
    Torr, Philip H. S.
    Zhang, Li
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 6877 - 6886
  • [37] Zhou Z-H., 2021, Machine learning, DOI [10.1007/978-981-15-1967-3_8, DOI 10.1007/978-981-15-1967-3, 10.1007/978-981-15-1967-313]
  • [38] Zisserman A., 2010, International journal of computer vision, DOI DOI 10.1007/s11263-009-0275-4