Towards Grand Unified Representation Learning for Unsupervised Visible-Infrared Person Re-Identification

被引:37
作者
Bin Yang [1 ]
Chen, Jun [1 ]
Ye, Mang [1 ]
机构
[1] Wuhan Univ, Sch Comp Sci, Natl Engn Res Ctr Multimedia Software, Hubei Luojia Lab, Wuhan, Peoples R China
来源
2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023) | 2023年
基金
中国国家自然科学基金;
关键词
D O I
10.1109/ICCV51070.2023.01016
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Unsupervised learning visible-infrared person re-identification (USL-VI-ReID) is an extremely important and challenging task, which can alleviate the issue of expensive cross-modality annotations. Existing works focus on handling the cross-modality discrepancy under unsupervised conditions. However, they ignore the fact that USL-VI-ReID is a cross-modality retrieval task with the hierarchical discrepancy, i.e., camera variation and modality discrepancy, resulting in clustering inconsistencies and ambiguous cross-modality label association. To address these issues, we propose a hierarchical framework to learn grand unified representation (GUR) for USL-VI-ReID. The grand unified representation lies in two aspects: 1) GUR adopts a bottom-up domain learning strategy with a cross- memory association embedding module to explore the information of hierarchical domains, i.e., intra-camera, inter-camera, and inter-modality domains, learning a unified and robust representation against hierarchical discrepancy. 2) To unify the identities of the two modalities, we develop a cross-modality label unification module that constructs a cross-modality affinity matrix as a bridge for propagating labels between two modalities. Then, we utilize the homogeneous structure matrix to smooth the propagated labels, ensuring that the label structure within one modality remains unchanged. Extensive experiments demonstrate that our GUR framework significantly outperforms existing USL-VI-ReID methods, and even surpasses some supervised counterparts.
引用
收藏
页码:11035 / 11045
页数:11
相关论文
共 59 条
[1]  
[Anonymous], 2022, Tribology Letters, DOI DOI 10.1007/S11249-022-01638-8
[2]  
[Anonymous], 2019, NIPS
[3]   Structure-Aware Positional Transformer for Visible-Infrared Person Re-Identification [J].
Chen, Cuiqun ;
Ye, Mang ;
Qi, Meibin ;
Wu, Jingjing ;
Jiang, Jianguo ;
Lin, Chia-Wen .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2022, 31 :2352-2364
[4]   ICE: Inter-instance Contrastive Encoding for Unsupervised Person Re-identification [J].
Chen, Hao ;
Lagadec, Benoit ;
Bremond, Francois .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :14940-14949
[5]   Joint Generative and Contrastive Learning for Unsupervised Person Re-identification [J].
Chen, Hao ;
Wang, Yaohui ;
Lagadec, Benoit ;
Dantcheva, Antitza ;
Bremond, Francois .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :2004-2013
[6]   Instance-Guided Context Rendering for Cross-Domain Person Re-Identification [J].
Chen, Yanbei ;
Zhu, Xiatian ;
Gong, Shaogang .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :232-242
[7]   Part-based Pseudo Label Refinement for Unsupervised Person Re-identification [J].
Cho, Yoonki ;
Kim, Woo Jae ;
Hong, Seunghoon ;
Yoon, Sung-Eui .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, :7298-7308
[8]   Hi-CMD: Hierarchical Cross-Modality Disentanglement for Visible-Infrared Person Re-Identification [J].
Choi, Seokeon ;
Lee, Sumin ;
Kim, Youngeun ;
Kim, Taekyung ;
Kim, Changick .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :10254-10263
[9]  
Dai Zuozhuo, 2021, ARXIV210311568
[10]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848