Visible-infrared person re-identification model based on feature consistency and modal indistinguishability

被引:7
作者
Sun, Jia [1 ]
Li, Yanfeng [1 ]
Chen, Houjin [1 ]
Peng, Yahui [1 ]
Zhu, Jinlei [1 ,2 ]
机构
[1] Beijing Jiaotong Univ, Sch Elect Informat Engn, Beijing 100044, Peoples R China
[2] Synth Elect Technol Co Ltd, Jinan, Peoples R China
基金
中国国家自然科学基金;
关键词
Person re-identification; Cross-modality; Feature consistency; Modal indistinguishability;
D O I
10.1007/s00138-022-01368-w
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Visible-infrared person re-identification (VI-ReID) is used to search person images across cameras under different modalities, which can address the limitation of visible-based ReID in dark environments. Intra-class discrepancy and feature-expression discrepancy caused by modal changes are two major problems in VI-ReID. To address these problems, a VI-ReID model based on feature consistency and modal indistinguishability is proposed. Specifically, image features of the two modalities are obtained through a one-stream network. Aiming at the problem of intra-class discrepancy, the class-level central consistency loss is developed, which makes the two modal feature centroids of the same identity close to their class centroids. To reduce the discrepancy of feature expression, a modal adversarial learning strategy is designed to distinguish whether the features are consistent with the modality attributes. The aim is that the generator generates the feature which the discriminator cannot distinguish its modality, thereby shortening the modality discrepancy and improving the modal indistinguishability. The generator network is optimized by the cross-entropy loss, the triplet loss, the proposed central consistency loss and the modal adversarial loss. The discriminator network is optimized by the modal adversarial loss. Experiments on the two datasets SYSU-MM01 and RegDB demonstrate that our method has better VI-ReID performance. The code has been released in .
引用
收藏
页数:14
相关论文
共 41 条
[1]   An efficient framework for visible-infrared cross modality person re-identification [J].
Basaran, Emrah ;
Gokmen, Muhittin ;
Kamasak, Mustafa E. .
SIGNAL PROCESSING-IMAGE COMMUNICATION, 2020, 87
[2]  
Chen T, 2020, Arxiv, DOI [arXiv:2002.05709, 10.48550/arXiv.2002.05709, DOI 10.48550/ARXIV.2002.05709]
[3]   Neural Feature Search for RGB-Infrared Person Re-Identification [J].
Chen, Yehansen ;
Wan, Lin ;
Li, Zhihang ;
Jing, Qianyan ;
Sun, Zongyuan .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :587-597
[4]   Hi-CMD: Hierarchical Cross-Modality Disentanglement for Visible-Infrared Person Re-Identification [J].
Choi, Seokeon ;
Lee, Sumin ;
Kim, Youngeun ;
Kim, Taekyung ;
Kim, Changick .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :10254-10263
[5]  
Dai PY, 2018, PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P677
[6]  
Dai ZZ, 2021, Arxiv, DOI [arXiv:2103.11568, 10.48550/arXiv.2103.11568]
[7]   Learning Modality-Specific Representations for Visible-Infrared Person Re-Identification [J].
Feng, Zhanxiang ;
Lai, Jianhuang ;
Xie, Xiaohua .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 :579-590
[8]   Multi-shot person re-identification based on appearance and spatial-temporal cues in a large camera network [J].
Frikha, Mayssa ;
Fendri, Emna ;
Hammami, Mohamed .
MACHINE VISION AND APPLICATIONS, 2021, 32 (04)
[9]  
Ge Y. X., 2020, C WORKSHOP NEURAL IN
[10]  
Goodfellow IJ, 2014, ADV NEUR IN, V27, P2672