Knowledge self-distillation for visible-infrared cross-modality person re-identification

被引:7
作者
Zhou, Yu [1 ]
Li, Rui [1 ]
Sun, Yanjing [1 ,2 ]
Dong, Kaiwen [1 ]
Li, Song [1 ,2 ]
机构
[1] China Univ Min & Technol, Sch Informat & Control Engn, Xuzhou 221116, Jiangsu, Peoples R China
[2] Xuzhou Engn Res Ctr Intelligent Ind Safety & Emer, Xuzhou 221116, Jiangsu, Peoples R China
基金
中国国家自然科学基金;
关键词
Visible-Infrared person Re-IDentification (VI-ReID); Cross-modality; Knowledge self-distillation (KSD); ATTENTION; ALIGNMENT;
D O I
10.1007/s10489-021-02814-4
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Visible-Infrared cross-modality person Re-IDentification (VI-ReID) is a tough task due to the large modality discrepancy and intra-modality variations. Even so, increasing interest has still been attracted by virtue of its significant role in public security. In this paper, we propose a novel VI-ReID method based on Knowledge Self-Distillation (KSD), which aims to improve the discrimination ability of the common neural network through better feature exploration. KSD is achieved by first constructing shallow recognizers with the same structure as the deepest recognizer in the same convolutional neural network and then using the deepest one to teach shallower ones under multi-dimensional supervision. Subsequently, the lower-level features extracted from shallower layers that have absorbed deep knowledge further boost the higher-level feature learning in turn. During the training process, multi-dimensional loss functions are integrated as the mentor for more effective learning supervision. Finally, a VI-ReID model with better feature representation capability is produced via abundant knowledge transfer and feedback. Extensive experiments on two public databases demonstrate the significant superiority of the proposed method in terms of identification accuracy. Furthermore, our method is also proved to be effective to achieve model lightweight on the premise of guaranteeing the performance, which indicates the huge application potential on resource-limited edge devices.
引用
收藏
页码:10617 / 10631
页数:15
相关论文
共 49 条
[1]   Beyond triplet loss: a deep quadruplet network for person re-identification [J].
Chen, Weihua ;
Chen, Xiaotang ;
Zhang, Jianguo ;
Huang, Kaiqi .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :1320-1329
[2]   Learning domain invariant and specific representation for cross-domain person re-identification [J].
Chong, Yanwen ;
Peng, Chengwei ;
Zhang, Chen ;
Wang, Yujie ;
Feng, Wenqiang ;
Pan, Shaoming .
APPLIED INTELLIGENCE, 2021, 51 (08) :5219-5232
[3]  
Dai PY, 2018, PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P677
[4]   Learning Modality-Specific Representations for Visible-Infrared Person Re-Identification [J].
Feng, Zhanxiang ;
Lai, Jianhuang ;
Xie, Xiaohua .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 :579-590
[5]   Learning View-Specific Deep Networks for Person Re-Identification [J].
Feng, Zhanxiang ;
Lai, Jianhuang ;
Xie, Xiaohua .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2018, 27 (07) :3472-3483
[6]  
Fu Y, 2019, AAAI CONF ARTIF INTE, P8287
[7]  
Hao Y, 2019, AAAI CONF ARTIF INTE, P8385
[8]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[9]  
Hermans A., 2017, ARXIV PREPRINT ARXIV
[10]   Searching for MobileNetV3 [J].
Howard, Andrew ;
Sandler, Mark ;
Chu, Grace ;
Chen, Liang-Chieh ;
Chen, Bo ;
Tan, Mingxing ;
Wang, Weijun ;
Zhu, Yukun ;
Pang, Ruoming ;
Vasudevan, Vijay ;
Le, Quoc V. ;
Adam, Hartwig .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :1314-1324