Foreground separation knowledge distillation for object detection

被引:0
作者
Li, Chao [1 ]
Liu, Rugui [1 ]
Quan, Zhe [1 ]
Hu, Pengpeng [2 ]
Sun, Jun [1 ]
机构
[1] Jiangnan Univ, Sch Artificial Intelligence & Comp Sci, Wuxi, Jiangsu, Peoples R China
[2] Coventry Univ, Ctr Computat Sci & Math Modelling, Coventry, England
基金
中国国家自然科学基金;
关键词
Knowledge distillation; Object detection; Foreground separation; Channel feature;
D O I
10.7717/peerj-cs.2485
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In recent years, deep learning models have become predominant methods for computer vision tasks, but the large computation and storage requirements of many models make them challenging to deploy on devices with limited resources. Knowledge distillation (KD) is a widely used approach for model compression. However, when applied in the object detection problems, the existing KD methods either directly applies the feature map or simply separate the foreground from the background by using a binary mask, aligning the attention between the teacher and the student models. Unfortunately, these methods either completely overlook or fail to thoroughly eliminate noise, resulting in unsatisfactory model accuracy for student models. To address this issue, we propose a foreground separation distillation (FSD) method in this paper. The FSD method enables student models to distinguish between foreground and background using Gaussian heatmaps, reducing irrelevant information in the learning process. Additionally, FSD also extracts the channel feature by converting the spatial feature maps into probabilistic forms to fully utilize the knowledge in each channel of a well-trained teacher. Experimental results demonstrate that the YOLOX detector enhanced with our distillation method achieved superior performance on both the fall detection and the VOC2007 datasets. For example, YOLOX with FSD achieved 73.1% mean average precision (mAP) on the Fall Detection dataset, which is 1.6% higher than the baseline. The code of FSD is accessible via https://doi.org/10.5281/zenodo.13829676.
引用
收藏
页数:22
相关论文
共 50 条
  • [41] KDNet: Leveraging Vision-Language Knowledge Distillation for Few-Shot Object Detection
    Ma, Mengyuan
    Qian, Lin
    Yin, Hujun
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING-ICANN 2024, PT II, 2024, 15017 : 153 - 167
  • [42] Voxel-to-Pillar: Knowledge Distillation of 3D Object Detection in Point Cloud
    Zhang, Jinbao
    Liu, Jun
    PROCEEDINGS OF THE 4TH EUROPEAN SYMPOSIUM ON SOFTWARE ENGINEERING, ESSE 2023, 2024, : 99 - 104
  • [43] Consistency- and dependence-guided knowledge distillation for object detection in remote sensing images
    Chen, Yixia
    Lin, Mingwei
    He, Zhu
    Polat, Kemal
    Alhudhaif, Adi
    Alenezi, Fayadh
    EXPERT SYSTEMS WITH APPLICATIONS, 2023, 229
  • [44] Task-balanced distillation for object detection
    Tang, Ruining
    Liu, Zhenyu
    Li, Yangguang
    Song, Yiguo
    Liu, Hui
    Wang, Qide
    Shao, Jing
    Duan, Guifang
    Tan, Jianrong
    PATTERN RECOGNITION, 2023, 137
  • [45] Foreground Gated Network for Surveillance Object Detection
    Fu, Zhihang
    Zhou, Chang
    Yong, Hongwei
    Jiang, Rongxin
    Tian, Xiang
    Chen, Yaowu
    Hua, Xian-Sheng
    2018 IEEE FOURTH INTERNATIONAL CONFERENCE ON MULTIMEDIA BIG DATA (BIGMM), 2018,
  • [46] CORRELATION AND FOREGROUND ATTENTION TO IMPROVE OBJECT DETECTION
    Dong, Yudi
    Yue, Xiaodong
    Xu, Zhikang
    Xie, Shaorong
    2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 3150 - 3154
  • [47] Knowledge distillation for object detection based on Inconsistency-based Feature Imitation and Global Relation Imitation
    Ju, Peng
    Zhang, Yi
    NEUROCOMPUTING, 2024, 566
  • [48] Multi-level knowledge distillation for low-resolution object detection and facial expression recognition
    Ma, Tingsong
    Tian, Wenhong
    Xie, Yuanlun
    KNOWLEDGE-BASED SYSTEMS, 2022, 240
  • [49] Foreground Object Detection by Motion-based Grouping of Object Parts
    Wang, Zhipeng
    Cui, Jinshi
    Zha, Hongbin
    Kagesawa, Masataka
    Ono, Shintaro
    Ikeuchi, Katsushi
    INTERNATIONAL JOURNAL OF INTELLIGENT TRANSPORTATION SYSTEMS RESEARCH, 2014, 12 (02) : 70 - 82
  • [50] MULTI-TEACHER DISTILLATION FOR INCREMENTAL OBJECT DETECTION
    Jiang, Le
    Cheng, Hongqiang
    Ye, Xiaozhou
    Ouyang, Ye
    2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, ICASSP 2024, 2024, : 5520 - 5524