Cross-Scale Transformer-Based Matching Network for Generalizable Person Re-Identification

被引:0
作者
Xiao, Junjie [1 ]
Jiang, Jinhua [2 ]
Huang, Jianyong [3 ]
Hu, Wei [1 ]
Zhang, Wenfeng [1 ]
机构
[1] Chongqing Normal Univ, Sch Comp & Informat Sci, Chongqing 401331, Peoples R China
[2] Nanjing Univ Aeronaut & Astronaut, Sch Mech & Elect Engn, Nanjing 210016, Peoples R China
[3] Qingdao Educ Equipment & Informat Technol Ctr, Qingdao 266022, Peoples R China
关键词
Transformers; Feature extraction; Convolution; Image matching; Standards; Pedestrians; Image segmentation; Identification of persons; Adaptation models; Accuracy; Domain generalization; person re-identification; deep image matching; cross-scale local respondence; deformable convolution;
D O I
10.1109/ACCESS.2025.3548321
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
While the person re-identification (Re-ID) task has made significant progress in closed-set setting in recent years, its generalizability to unknown domains continues to be limited. To tackle the issue, the domain generalization (DG) Re-ID task has been proposed. The current state-of-the-art approach involves deep feature matching, where key regions of image pairs are matched at the same scale. However, the method does not take into account the variability of angles in real image acquisition. To resolve the problem, we propose an innovative deep image matching framework called Cross-scale Transformer-based Matching Network (CTMN) for DG Re-ID task. CTMN model matches two images through cross-scale local respondence rather than using fixed representations. The Transformer is specifically adjusted to enable effective local interactions between query and gallery images across different scales. Additionally, deformable convolution is incorporated to better segment the local regions of the person before the procedure for matching image pairs. Lastly, the Style Normalization Module (SNM) is added to remove identity-irrelevant features, improving the matching results. Extensive experiments on multiple DG Re-ID tasks demonstrate the advantages of our proposed method over existing state-of-the-arts.
引用
收藏
页码:47406 / 47417
页数:12
相关论文
共 57 条
[1]  
Arjovsky M, 2020, Arxiv, DOI arXiv:1907.02893
[2]  
Cha J, 2022, LECT NOTES COMPUT SC, V13683, P440
[3]   Meta Batch-Instance Normalization for Generalizable Person Re-Identification [J].
Choi, Seokeon ;
Kim, Taekyung ;
Jeong, Minki ;
Park, Hyoungseob ;
Kim, Changick .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :3424-3434
[4]  
Christlein Vincent, 2019, 2019 International Conference on Document Analysis and Recognition (ICDAR). Proceedings, P1090, DOI 10.1109/ICDAR.2019.00177
[5]   Bridging the Source-to-Target Gap for Cross-Domain Person Re-identification with Intermediate Domains [J].
Dai, Yongxing ;
Sun, Yifan ;
Liu, Jun ;
Tong, Zekun ;
Duan, Ling-Yu .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2025, 133 (01) :410-434
[6]   Generalizable Person Re-identification with Relevance-aware Mixture of Experts [J].
Dai, Yongxing ;
Li, Xiaotong ;
Liu, Jun ;
Tong, Zekun ;
Duan, Ling-Yu .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :16140-16149
[7]  
Dosovitskiy A, 2021, Arxiv, DOI arXiv:2010.11929
[8]  
Dou Q, 2019, ADV NEUR IN, V32
[9]   Adaptive Methods for Real-World Domain Generalization [J].
Dubey, Abhimanyu ;
Ramanathan, Vignesh ;
Pentland, Alex ;
Mahajan, Dhruv .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :14335-14344
[10]  
Ganin Y, 2016, J MACH LEARN RES, V17