Knowledge Distillation Meets Open-Set Semi-supervised Learning

被引:2
|
作者
Yang, Jing [1 ]
Zhu, Xiatian [2 ,3 ]
Bulat, Adrian [2 ]
Martinez, Brais [2 ]
Tzimiropoulos, Georgios [2 ,4 ]
机构
[1] Univ Nottingham, Nottingham, England
[2] Samsung AI Ctr, Cambridge, England
[3] Univ Surrey, Guildford, England
[4] Queen Mary Univ London, London, England
关键词
Knowledge distillation; Structured representational knowledge; Open-set semi-supervised learning; Out-of-distribution;
D O I
10.1007/s11263-024-02192-7
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Existing knowledge distillation methods mostly focus on distillation of teacher's prediction and intermediate activation. However, the structured representation, which arguably is one of the most critical ingredients of deep models, is largely overlooked. In this work, we propose a novel semantic representational distillation (SRD) method dedicated for distilling representational knowledge semantically from a pretrained teacher to a target student. The key idea is that we leverage the teacher's classifier as a semantic critic for evaluating the representations of both teacher and student and distilling the semantic knowledge with high-order structured information over all feature dimensions. This is accomplished by introducing a notion of cross-network logit computed through passing student's representation into teacher's classifier. Further, considering the set of seen classes as a basis for the semantic space in a combinatorial perspective, we scale SRD to unseen classes for enabling effective exploitation of largely available, arbitrary unlabeled training data. At the problem level, this establishes an interesting connection between knowledge distillation with open-set semi-supervised learning (SSL). Extensive experiments show that our SRD outperforms significantly previous state-of-the-art knowledge distillation methods on both coarse object classification and fine face recognition tasks, as well as less studied yet practically crucial binary network distillation. Under more realistic open-set SSL settings we introduce, we reveal that knowledge distillation is generally more effective than existing out-of-distribution sample detection, and our proposed SRD is superior over both previous distillation and SSL competitors. The source code is available at https://github.com/jingyang2017/SRD_ossl.
引用
收藏
页码:315 / 334
页数:20
相关论文
共 50 条
  • [31] ASSET: A Semi-supervised Approach for Entity Typing in Knowledge Graphs
    Zahera, Hamada M.
    Heindorf, Stefan
    Ngomo, Axel-Cyrille Ngonga
    PROCEEDINGS OF THE 11TH KNOWLEDGE CAPTURE CONFERENCE (K-CAP '21), 2021, : 261 - 264
  • [32] Semi-supervised knowledge distillation framework for global-scale urban man-made object remote sensing mapping
    Chen, Dingyuan
    Ma, Ailong
    Zhong, Yanfei
    INTERNATIONAL JOURNAL OF APPLIED EARTH OBSERVATION AND GEOINFORMATION, 2023, 122
  • [33] Toward generalizable robot vision guidance in real-world operational manufacturing factories: A Semi-Supervised Knowledge Distillation approach
    Zhao, Zizhou
    Lyu, Junyu
    Chu, Yinghao
    Liu, Ke
    Cao, Daofan
    Wu, Changning
    Qin, Longjun
    Qin, Shiwei
    ROBOTICS AND COMPUTER-INTEGRATED MANUFACTURING, 2024, 86
  • [34] Safe semi-supervised learning using a bayesian neural network
    Bae, Jinsoo
    Lee, Minjung
    Kim, Seoung Bum
    INFORMATION SCIENCES, 2022, 612 : 453 - 464
  • [35] Bootstrapping Semi-supervised Medical Image Segmentation with Anatomical-Aware Contrastive Distillation
    You, Chenyu
    Dai, Weicheng
    Min, Yifei
    Staib, Lawrence
    Duncan, James S.
    INFORMATION PROCESSING IN MEDICAL IMAGING, IPMI 2023, 2023, 13939 : 641 - 653
  • [36] OUT-OF-DISTRIBUTION AS A TARGET CLASS IN SEMI-SUPERVISED LEARNING
    Tadros, Antoine
    Drouyer, Sebastien
    von Gioi, Rafael Grompone
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 3249 - 3252
  • [37] Self-supervised knowledge distillation for complementary label learning
    Liu, Jiabin
    Li, Biao
    Lei, Minglong
    Shi, Yong
    NEURAL NETWORKS, 2022, 155 : 318 - 327
  • [38] FedKD-IDS: A robust intrusion detection system using knowledge distillation-based semi-supervised federated learning and anti-poisoning attack mechanism
    Quyen, Nguyen Huu
    Duy, Phan The
    Nguyen, Ngo Thao
    Khoa, Nghi Hoang
    Pham, Van-Hau
    INFORMATION FUSION, 2025, 117
  • [39] SEMI-SUPERVISED LEARNING WITH OUT-OF-DISTRIBUTION UNLABELED SAMPLES FOR RETINAL IMAGE CLASSIFICATION
    Jia, Lize
    Guo, Jia
    Zhang, Weihang
    Liu, Hanruo
    Wang, Ningli
    Li, Huiqi
    2023 IEEE 20TH INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING, ISBI, 2023,
  • [40] A Semi-supervised Abdominal Multi-organ Pan-Cancer Segmentation Framework with Knowledge Distillation and Multi-label Fusion
    Zhang, Zengmin
    Duan, Xiaomeng
    Peng, Yanjun
    Li, Zhengyu
    FAST, LOW-RESOURCE, AND ACCURATE ORGAN AND PAN-CANCER SEGMENTATION IN ABDOMEN CT, FLARE 2023, 2024, 14544 : 346 - 361