Dual-Stage Attribute Embedding and Modality Consistency Learning-Based Visible-Infrared Person Re-Identification

被引:1
作者
Cheng, Zhuxuan [1 ,2 ,3 ]
Fan, Huijie [2 ,3 ]
Wang, Qiang [4 ]
Liu, Shiben [2 ,3 ]
Tang, Yandong [2 ,3 ]
机构
[1] Shenyang Univ Chem Technol, Sch Informat Engn, Shenyang 110142, Peoples R China
[2] Chinese Acad Sci, Shenyang Inst Automat, State Key Lab Robot, Shenyang 110016, Peoples R China
[3] Chinese Acad Sci, Inst Robot & Intelligent Mfg, Shenyang 110003, Peoples R China
[4] Shenyang Univ, Key Lab Mfg Ind Integrated, Shenyang 110044, Peoples R China
基金
中国国家自然科学基金;
关键词
visible-infrared person re-identification; attribute information; dual-stage embedding; modality consistency; TRANSFORMER;
D O I
10.3390/electronics12244892
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Visible-infrared person re-identification (VI-ReID) is an emerging technology for realizing all-weather smart surveillance systems. To address the problem of pedestrian discriminative information being difficult to obtain and easy to lose, as well as the wide modality difference in the VI-ReID task, in this paper we propose a two-stage attribute embedding and modality consistency learning-based VI-ReID method. First, the attribute information embedding module introduces the fine-grained pedestrian information in the attribute label into the transformer backbone, enabling the backbone to extract identity-discriminative pedestrian features. After obtaining the pedestrian features, the attribute embedding enhancement module is utilized to realize the second-stage attribute information embedding, which reduces the adverse effect of losing the person discriminative information due to the deepening of network. Finally, the modality consistency learning loss is designed for constraining the network to mine the consistency information between two modalities in order to reduce the impact of modality difference on the recognition results. The results show that our method reaches 74.57% mAP on the SYSU-MM01 dataset in All Search mode and 87.02% mAP on the RegDB dataset in IR-to-VIS mode, with a performance improvement of 6.00% and 2.56%, respectively, proving that our proposed method is able to reach optimal performance compared to existing state-of-the-art methods.
引用
收藏
页数:16
相关论文
共 40 条
[1]   Structure-Aware Positional Transformer for Visible-Infrared Person Re-Identification [J].
Chen, Cuiqun ;
Ye, Mang ;
Qi, Meibin ;
Wu, Jingjing ;
Jiang, Jianguo ;
Lin, Chia-Wen .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2022, 31 :2352-2364
[2]   Neural Feature Search for RGB-Infrared Person Re-Identification [J].
Chen, Yehansen ;
Wan, Lin ;
Li, Zhihang ;
Jing, Qianyan ;
Sun, Zongyuan .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :587-597
[3]   Hi-CMD: Hierarchical Cross-Modality Disentanglement for Visible-Infrared Person Re-Identification [J].
Choi, Seokeon ;
Lee, Sumin ;
Kim, Youngeun ;
Kim, Taekyung ;
Kim, Changick .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :10254-10263
[4]  
Dai PY, 2018, PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P677
[5]   MEVID: Multi-view Extended Videos with Identities for Video Person Re-Identification [J].
Davila, Daniel ;
Du, Dawei ;
Lewis, Bryon ;
Funk, Christopher ;
Van Pelt, Joseph ;
Collins, Roderic ;
Corona, Kellie ;
Brown, Matt ;
McCloskey, Scott ;
Hoogs, Anthony ;
Clipp, Brian .
2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2023, :1634-1643
[6]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[7]  
Dosovitskiy A., 2021, P 9 INT C LEARN REPR, P2021
[8]   Learning Modality-Specific Representations for Visible-Infrared Person Re-Identification [J].
Feng, Zhanxiang ;
Lai, Jianhuang ;
Xie, Xiaohua .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 :579-590
[9]   CM-NAS: Cross-Modality Neural Architecture Search for Visible-Infrared Person Re-Identification [J].
Fu, Chaoyou ;
Hu, Yibo ;
Wu, Xiang ;
Shi, Hailin ;
Mei, Tao ;
He, Ran .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :11803-11812
[10]  
Ge Y., 2020, P INT C LEARN REPR A