Multi-Scale Explicit Matching and Mutual Subject Teacher Learning for Generalizable Person Re-Identification

被引:1
作者
Chen, Kaixiang [1 ]
Fang, Pengfei [2 ]
Ye, Zi [1 ]
Zhang, Liyan [1 ]
机构
[1] Nanjing Univ Aeronaut & Astronaut, Coll Comp Sci & Technol, Nanjing 211106, Peoples R China
[2] Southeast Univ, Sch Comp Sci & Engn, Nanjing 210096, Peoples R China
基金
中国国家自然科学基金;
关键词
Person re-identification; domain generalization; multi-scale; mutual-teacher;
D O I
10.1109/TCSVT.2024.3382322
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Domain generalization in person re-identification (DG-ReID) stands out as the most challenging task and practically important branch in the ReID field, which enables the direct deployment of pre-trained models in unseen and real scenarios. Recent works have made significant efforts in this task via the image-matching paradigm, which searches for the local correspondences in the feature maps. A common practice of employing pixel-wise matching is typically used to ensure efficient matching. This, however, makes the matching susceptible to deviations caused by identity-irrelevant pixel features. On the other hand, patch-wise matching also demonstrates that it will disregard the spatial orientation of pedestrians and amplify the impact of noise. To address the mentioned issues, this paper proposes the Multi-Scale Query-Adaptive Convolution (QAConv-MS) framework, which encodes patches in the feature maps to pixels using template kernels of various scales. This enables the matching process to enjoy broader receptive fields and robustness to orientations and noises. To stabilize the matching process and facilitate the independent learning of each sub-kernel within the template kernels to capture diverse local patterns, we propose the OrthoGonal Norm (OGNorm), which consists of two orthogonal normalizations. We also present Mutual Subject Teacher Learning (MSTL) to address the potential issues of overconfidence and overfitting in the model. MSTL allows two models to individually select the most challenging data for training, resulting in more dependable soft labels that can provide mutual supervision. Extensive experiments conducted in both single-source and multi-source setups offer compelling evidence of our framework's generalization and competitiveness.
引用
收藏
页码:8881 / 8895
页数:15
相关论文
共 71 条
[41]  
Xu B., Liang J., He L., Sun Z., Mimic embedding via adaptive aggregation: Learning generalizable person re-identification, Proc. Eur. Conf. Comput. Vis, pp. 372-388, (2022)
[42]  
Finn C., Abbeel P., Levine S., Model-agnostic meta-learning for fast adaptation of deep networks, Proc. Int. Conf. Mach. Learn., pp. 1126-1135, (2017)
[43]  
Zheng L., Shen L., Tian L., Wang S., Wang J., Tian Q., Scalable person re-identification: A benchmark, Proc. IEEE Int. Conf. Comput. Vis. (ICCV), pp. 1116-1124, (2015)
[44]  
Zheng Z., Zheng L., Yang Y., Unlabeled samples generated by GAN improve the person re-identification baseline in vitro, Proc. IEEE Int. Conf. Comput. Vis. (ICCV), pp. 3754-3762, (2017)
[45]  
Li W., Zhao R., Xiao T., Wang X., DeepReID: Deep filter pairing neural network for person re-identification, Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 152-159, (2014)
[46]  
Xiao T., Li S., Wang B., Lin L., Wang X., Joint detection and identification feature learning for person search, (2016)
[47]  
Wei L., Zhang S., Gao W., Tian Q., Person transfer GAN to bridge domain gap for person re-identification, Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., pp. 79-88, (2018)
[48]  
Wang Y., Liao S., Shao L., Surpassing real-world source training data: Random 3D characters for generalizable person re-identification, Proc. 28th ACM Int. Conf. Multimedia, pp. 3422-3430, (2020)
[49]  
Zhuang Z., Wei L., Xie L., Ai H., Tian Q., Camera-based batch normalization: An effective distribution alignment method for person reidentification, IEEE Trans. Circuits Syst. Video Technol., 32, 1, pp. 374-387, (2022)
[50]  
Ni H., Li Y., Gao L., Shen H.T., Song J., Part-aware transformer for generalizable person re-identification, Proc. IEEE/CVF Int. Conf. Comput. Vis. (ICCV), pp. 11280-11289, (2023)