G2DA: Geometry-guided dual-alignment learning for RGB-infrared person re-identification

被引:44
作者
Wan, Lin [1 ]
Sun, Zongyuan [1 ]
Jing, Qianyan [1 ]
Chen, Yehansen [1 ]
Lu, Lijing [2 ]
Li, Zhihang [2 ]
机构
[1] China Univ Geosci, Sch Comp Sci, Wuhan, Peoples R China
[2] Univ Chinese Acad Sci, Sch Artificial Intelligence, Beijing, Peoples R China
关键词
Person re-identification; Cross-modality matching; optimal transport; Feature alignment; Channel exchange; INVARIANT; NETWORK;
D O I
10.1016/j.patcog.2022.109150
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
RGB-Infrared (IR) person re-identification aims to retrieve person-of-interest from heterogeneous cam-eras, easily suffering from large image modality discrepancy caused by different sensing wavelength ranges. Existing works usually minimize such discrepancy by aligning modality distribution of global features, while neglecting deep semantics and high-order structural relations within each class. This might render the misalignment between heterogeneous samples. In this paper, we propose Geometry -Guided Dual-Alignment (G2DA) learning, which yields better sample-level modality alignment for RGB-IR ReID by solving a graph-enabled distribution matching task that maximizes agreement between multi -modality node representations considering edge topology. Specifically, we covert RGB/IR images into semantic-aligned graphs, in which whole-part features and their similarities are represented by nodes and associated edges, respectively. To simultaneously implement node-and edge-wise alignment (Dual Alignment), we introduce Optimal Transport (OT) as a metric to calculate cross-modality human body matching scores. By minimizing the displacement cost across RGB-IR graphs, G2DA could learn not just modality-invariant but structurally consistent cross-modality representations. Furthermore, we advance a Message Fusion Attention (MFA) mechanism to adaptively smooth the node representations within each RGB/IR graph, effectively alleviating occlusions caused by other individuals and/or objects. Extensive ex-periments on two standard benchmark datasets validate the superiority of G2DA, yielding competitive performance against previous state-of-the-arts.(c) 2022 Elsevier Ltd. All rights reserved.
引用
收藏
页数:12
相关论文
共 58 条
[1]   Deep-Person: Learning discriminative deep features for person Re-Identification [J].
Bai, Xiang ;
Yang, Mingkun ;
Huang, Tengteng ;
Dou, Zhiyong ;
Yu, Rui ;
Xu, Yongchao .
PATTERN RECOGNITION, 2020, 98
[2]  
Chen Liqun, 2019, 25 AMERICAS C INFORM
[3]   Neural Feature Search for RGB-Infrared Person Re-Identification [J].
Chen, Yehansen ;
Wan, Lin ;
Li, Zhihang ;
Jing, Qianyan ;
Sun, Zongyuan .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :587-597
[4]   Hi-CMD: Hierarchical Cross-Modality Disentanglement for Visible-Infrared Person Re-Identification [J].
Choi, Seokeon ;
Lee, Sumin ;
Kim, Youngeun ;
Kim, Taekyung ;
Kim, Changick .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :10254-10263
[5]   Optimal Transport for Domain Adaptation [J].
Courty, Nicolas ;
Flamary, Remi ;
Tuia, Devis ;
Rakotomamonjy, Alain .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (09) :1853-1865
[6]  
Dai PY, 2018, PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P677
[7]   Learning Modality-Specific Representations for Visible-Infrared Person Re-Identification [J].
Feng, Zhanxiang ;
Lai, Jianhuang ;
Xie, Xiaohua .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 :579-590
[8]   CM-NAS: Cross-Modality Neural Architecture Search for Visible-Infrared Person Re-Identification [J].
Fu, Chaoyou ;
Hu, Yibo ;
Wu, Xiang ;
Shi, Hailin ;
Mei, Tao ;
He, Ran .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :11803-11812
[9]   Modality adversarial neural network for visible-thermal person re-identification [J].
Hao, Yi ;
Li, Jie ;
Wang, Nannan ;
Gao, Xinbo .
PATTERN RECOGNITION, 2020, 107
[10]  
Hao Y, 2019, AAAI CONF ARTIF INTE, P8385