The Staged Knowledge Distillation in Video Classification: Harmonizing Student Progress by a Complementary Weakly Supervised Framework

被引:2
|
作者
Wang, Chao [1 ]
Tang, Zheng [2 ]
机构
[1] China Acad Railway Sci, Beijing 100081, Peoples R China
[2] NVIDIA, Redmond, WA 98052 USA
关键词
Training; Uncertainty; Correlation; Generators; Data models; Task analysis; Computational modeling; Knowledge distillation; weakly supervised learning; teacher-student architecture; substage learning process; video classification; label-efficient learning;
D O I
10.1109/TCSVT.2023.3294977
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In the context of label-efficient learning on video data, the distillation method and the structural design of the teacher-student architecture have a significant impact on knowledge distillation. However, the relationship between these factors has been overlooked in previous research. To address this gap, we propose a new weakly supervised learning framework for knowledge distillation in video classification that is designed to improve the efficiency and accuracy of the student model. Our approach leverages the concept of substage-based learning to distill knowledge based on the combination of student substages and the correlation of corresponding substages. We also employ the progressive cascade training method to address the accuracy loss caused by the large capacity gap between the teacher and the student. Additionally, we propose a pseudo-label optimization strategy to improve the initial data label. To optimize the loss functions of different distillation substages during the training process, we introduce a new loss method based on feature distribution. We conduct extensive experiments on both real and simulated data sets, demonstrating that our proposed approach outperforms existing distillation methods in terms of knowledge distillation for video classification tasks. Our proposed substage-based distillation approach has the potential to inform future research on label-efficient learning for video data.
引用
收藏
页码:6646 / 6660
页数:15
相关论文
共 50 条
  • [21] A Novel Knowledge Distillation Method for Self-Supervised Hyperspectral Image Classification
    Chi, Qiang
    Lv, Guohua
    Zhao, Guixin
    Dong, Xiangjun
    REMOTE SENSING, 2022, 14 (18)
  • [22] Weakly supervised object localization via knowledge distillation based on foreground-background contrast
    Ma, Siteng
    Hou, Biao
    Li, Zhihao
    Wu, Zitong
    Guo, Xianpeng
    Yang, Chen
    Jiao, Licheng
    NEUROCOMPUTING, 2024, 576
  • [23] A Memory-Assisted Knowledge Transferring Framework with Curriculum Anticipation for Weakly Supervised Online Activity Detection
    Liu, Tianshan
    Lam, Kin-Man
    Bao, Bing-Kun
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2024, : 1940 - 1963
  • [24] Cross-Video Contextual Knowledge Exploration and Exploitation for Ambiguity Reduction in Weakly Supervised Temporal Action Localization
    Zhang, Songchun
    Zhao, Chunhui
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (06) : 4568 - 4580
  • [25] SIAVC: Semi-Supervised Framework for Industrial Accident Video Classification
    Li, Zuoyong
    Lin, Qinghua
    Fan, Haoyi
    Zhao, Tiesong
    Zhang, David
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2025, 35 (03) : 2603 - 2615
  • [26] Heterogeneous Student Knowledge Distillation From BERT Using a Lightweight Ensemble Framework
    Lin, Ching-Sheng
    Tsai, Chung-Nan
    Jwo, Jung-Sing
    Lee, Cheng-Hsiung
    Wang, Xin
    IEEE ACCESS, 2024, 12 : 33079 - 33088
  • [27] Collaborative deep semi-supervised learning with knowledge distillation for surface defect classification
    Manivannan, Siyamalan
    COMPUTERS & INDUSTRIAL ENGINEERING, 2023, 186
  • [28] A coded knowledge distillation framework for image classification based on adaptive JPEG encoding
    Salamah, Ahmed H.
    Hamidi, Shayan Mohajer
    Yang, En-Hui
    PATTERN RECOGNITION, 2025, 158
  • [29] A Multi-teacher Knowledge Distillation Framework for Distantly Supervised Relation Extraction with Flexible Temperature
    Fei, Hongxiao
    Tan, Yangying
    Huang, Wenti
    Long, Jun
    Huang, Jincai
    Yang, Liu
    WEB AND BIG DATA, PT II, APWEB-WAIM 2023, 2024, 14332 : 103 - 116
  • [30] A Teacher-Student Knowledge Distillation Framework for Enhanced Detection of Anomalous User Activity
    Hsu, Chan
    Ku, Chan-Tung
    Wang, Yuwen
    Hsieh, Minchen
    Wu, Jun-Ting
    Hsieh, Yunhsiang
    Chang, PoFeng
    Lu, Yimin
    Kang, Yihuang
    2023 IEEE 24TH INTERNATIONAL CONFERENCE ON INFORMATION REUSE AND INTEGRATION FOR DATA SCIENCE, IRI, 2023, : 20 - 21