Learning to Reduce Dual-level Discrepancy for Infrared-Visible Person Re-identification

被引:373
作者
Wang, Zhixiang [1 ]
Wang, Zheng [2 ]
Zheng, Yinqiang [2 ]
Chuang, Yung-Yu [1 ]
Satoh, Shin'ichi [2 ,3 ]
机构
[1] Natl Taiwan Univ, Taipei, Taiwan
[2] Natl Inst Informat, Tokyo, Japan
[3] Univ Tokyo, Tokyo, Japan
来源
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019) | 2019年
关键词
D O I
10.1109/CVPR.2019.00071
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Infrared-Visible person RE-IDentification (IV-REID) is a rising task. Compared to conventional person re-identification (re-ID), IV-REID concerns the additional modality discrepancy originated from the different imaging processes of spectrum cameras, in addition to the person's appearance discrepancy caused by viewpoint changes, pose variations and deformations presented in the conventional re-ID task. The co-existed discrepancies make IV-REID more difficult to solve. Previous methods attempt to reduce the appearance and modality discrepancies simultaneously using feature-level constraints. It is however difficult to eliminate the mixed discrepancies using only feature-level constraints. To address the problem, this paper introduces a novel Dual-level Discrepancy Reduction Learning ((DRL)-R-2) scheme which handles the two discrepancies separately. For reducing the modality discrepancy, an image-level sub-network is trained to translate an infrared image into its visible counterpart and a visible image to its infrared version. With the image-level sub-network, we can unify the representations for images with different modalities. With the help of the unified multi-spectral images, a feature-level sub-network is trained to reduce the remaining appearance discrepancy through feature embedding. By cascading the two sub-networks and training them jointly, the dual-level reductions take their responsibilities cooperatively and attentively. Extensive experiments demonstrate the proposed approach outperforms the state-of-the-art methods.
引用
收藏
页码:618 / 626
页数:9
相关论文
共 29 条
[1]  
[Anonymous], P C COMP VIS PATT RE
[2]  
Chen Yu-Sheng, 2018, P C COMP VIS PATT RE
[3]  
Dai PY, 2018, PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P677
[4]  
Deng Weijian, 2018, P C COMP VIS PATT RE
[5]  
Kingma DP, 2014, ARXIV
[6]  
Li A, 2018, P C COMP VIS PATT RE
[7]  
Li Xia, 2018, P EUR C COMP VIS
[8]  
Liao Shengcai, 2015, P INT C COMP VIS
[9]  
Liao Shengcai, 2015, P C COMP VIS PATT RE
[10]   Cross-Domain Visual Matching via Generalized Similarity Measure and Feature Learning [J].
Lin, Liang ;
Wang, Guangrun ;
Zuo, Wangmeng ;
Feng, Xiangchu ;
Zhang, Lei .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (06) :1089-1102