Compressing the Multiobject Tracking Model via Knowledge Distillation

被引:4
作者
Liang, Tianyi [1 ]
Wang, Mengzhu [2 ]
Chen, Junyang [3 ]
Chen, Dingyao [4 ]
Luo, Zhigang [4 ]
Leung, Victor C. M. [3 ]
机构
[1] Inspur Grp Co Ltd, Jinan 250101, Shandong, Peoples R China
[2] DAMO Acad, Alibaba Grp, Hangzhou, Peoples R China
[3] Shenzhen Univ, Coll Comp Sci & Software Engn, Shenzhen, Peoples R China
[4] Natl Univ Def Technol, Coll Comp, Changsha, Peoples R China
基金
中国国家自然科学基金;
关键词
Knowledge distillation (KD); model compression; multiobject tracking (MOT); MULTITARGET;
D O I
10.1109/TCSS.2023.3293882
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Recent multiobject tracking (MOT) methods usually use very deep neural networks to achieve competitive accuracy, which inevitably results in degraded inference speed. To strike a better balance between tracking accuracy and speed, in this work, we propose to compress the MOT model via knowledge distillation (KD), enabling the more lightweight student model to obtain similar performance as the teacher model. Nonetheless, despite KD has been well studied for simpler tasks such as image classification, the complexity of MOT poses new challenges because the MOT model is more sensitive to foreground information than the classification model. To deal with that, we first propose attention-guided feature distillation, which focuses the student model on the crucial region (foreground and the region with strong discrepancy against itself) of the teacher's feature map. Moreover, we propose foreground mask, which leverages the knowledge from the teacher model to filter out the low-quality soft labels from the background, thereby reducing their negative effects for distillation. Evaluations on several benchmarks demonstrate that the proposed KD method can make the student network achieve leading performance, meanwhile running faster than the teacher network 20.0%-27.4% and reducing the parameters 28.5%-87.1%. To the best of our knowledge, this is the first work to compress the MOT model via KD.
引用
收藏
页码:2713 / 2723
页数:11
相关论文
共 67 条
  • [41] Multiobject Tracking by Submodular Optimization
    Shen, Jianbing
    Liang, Zhiyuan
    Liu, Jianhong
    Sun, Hanqiu
    Shao, Ling
    Tao, Dacheng
    [J]. IEEE TRANSACTIONS ON CYBERNETICS, 2019, 49 (06) : 1990 - 2001
  • [42] Sun P., 2020, ARXIV
  • [43] Deep Affinity Network for Multiple Object Tracking
    Sun, Shijie
    Akhtar, Naveed
    Song, HuanSheng
    Mian, Ajmal S.
    Shah, Mubarak
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2021, 43 (01) : 104 - 119
  • [44] Focal Loss for Dense Object Detection
    Lin, Tsung-Yi
    Goyal, Priya
    Girshick, Ross
    He, Kaiming
    Dollar, Piotr
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 2999 - 3007
  • [45] Distilling Object Detectors With Fine-Grained Feature Imitation
    Wang, Tao
    Yuan, Li
    Zhang, Xiaopeng
    Feng, Jiashi
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 4928 - 4937
  • [46] Wang Wenhui, 2020, Advances in Neural Information Processing Systems, V33
  • [47] Wang Z., 2020, P EUROPEAN C COMPUT, P107, DOI [10.1007/978-3-030-58621-8_7, DOI 10.1007/978-3-030-58621-8_7, DOI 10.1007/978-3-030-58621-87]
  • [48] Wojke N, 2017, IEEE IMAGE PROC, P3645, DOI 10.1109/ICIP.2017.8296962
  • [49] CBAM: Convolutional Block Attention Module
    Woo, Sanghyun
    Park, Jongchan
    Lee, Joon-Young
    Kweon, In So
    [J]. COMPUTER VISION - ECCV 2018, PT VII, 2018, 11211 : 3 - 19
  • [50] Track to Detect and Segment: An Online Multi-Object Tracker
    Wu, Jialian
    Cao, Jiale
    Song, Liangchen
    Wang, Yu
    Yang, Ming
    Yuan, Junsong
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 12347 - 12356