Learning Sparse and Identity-Preserved Hidden Attributes for Person Re-Identification

被引:71
作者
Wang, Zheng [1 ]
Jiang, Junjun [2 ,3 ]
Wu, Yang [4 ]
Ye, Mang [5 ]
Bai, Xiang [6 ]
Satoh, Shin'ichi [1 ]
机构
[1] Natl Inst Informat, Digital Content & Media Sci Res Div, Tokyo 1018430, Japan
[2] Harbin Inst Technol, Sch Comp Sci & Technol, Harbin 150001, Peoples R China
[3] Peng Cheng Lab, Shenzhen 518066, Peoples R China
[4] Nara Inst Sci & Technol, Int Collaborat Lab Robot Vis, Inst Res Initiat, Nara 6300192, Japan
[5] Incept Inst Artificial Intelligence, Abu Dhabi, U Arab Emirates
[6] Huazhong Univ Sci & Technol, Sch Elect Informat & Commun, Wuhan 430074, Peoples R China
关键词
Semantics; Deep learning; Visualization; Feature extraction; Image reconstruction; Clothing; Training; Person re-identification; attribute learning; generation; discrimination; NETWORK;
D O I
10.1109/TIP.2019.2946975
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Person re-identification (Re-ID) aims at matching person images captured in non-overlapping camera views. To represent person appearance, low-level visual features are sensitive to environmental changes, while high-level semantic attributes, such as "short-hair" or "long-hair", are relatively stable. Hence, researches have started to design semantic attributes to reduce the visual ambiguity. However, to train a prediction model for semantic attributes, it requires plenty of annotations, which are hard to obtain in practical large-scale applications. To alleviate the reliance on annotation efforts, we propose to incrementally generate Deep Hidden Attribute (DHA) based on baseline deep network for newly uncovered annotations. In particular, we propose an auto-encoder model that can be plugged into any deep network to mine latent information in an unsupervised manner. To optimize the effectiveness of DHA, we reform the auto-encoder model with additional orthogonal generation module, along with identity-preserving and sparsity constraints. 1) Orthogonally generating: In order to make DHAs different from each other, Singular Vector Decomposition (SVD) is introduced to generate DHAs orthogonally. 2) Identity-preserving constraint: The generated DHAs should be distinct for telling different persons, so we associate DHAs with person identities. 3) Sparsity constraint: To enhance the discriminability of DHAs, we also introduce the sparsity constraint to restrict the number of effective DHAs for each person. Experiments conducted on public datasets have validated the effectiveness of the proposed network. On two large-scale datasets, i.e., Market-1501 and DukeMTMC-reID, the proposed method outperforms the state-of-the-art methods.
引用
收藏
页码:2013 / 2025
页数:13
相关论文
共 59 条
[1]  
Abadi M, 2016, ACM SIGPLAN NOTICES, V51, P1, DOI [10.1145/2951913.2976746, 10.1145/3022670.2976746]
[2]  
[Anonymous], 2016, ARXIV
[3]   Scalable Person Re-identification on Supervised Smoothed Manifold [J].
Bai, Song ;
Bai, Xiang ;
Tian, Qi .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :3356-3365
[4]   Similarity Learning with Spatial Constraints for Person Re-identification [J].
Chen, Dapeng ;
Yuan, Zejian ;
Chen, Badong ;
Zheng, Nanning .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :1268-1277
[5]   Beyond triplet loss: a deep quadruplet network for person re-identification [J].
Chen, Weihua ;
Chen, Xiaotang ;
Zhang, Jianguo ;
Huang, Kaiqi .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :1320-1329
[6]  
Chen YC, 2015, PROCEEDINGS OF THE TWENTY-FOURTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE (IJCAI), P3402
[7]   Deep feature learning via structured graph Laplacian embedding for person re-identification [J].
Cheng, De ;
Gong, Yihong ;
Chang, Xiaojun ;
Shi, Weiwei ;
Hauptmann, Alexander ;
Zheng, Nanning .
PATTERN RECOGNITION, 2018, 82 :94-104
[8]   Person Re-Identification by Multi-Channel Parts-Based CNN with Improved Triplet Loss Function [J].
Cheng, De ;
Gong, Yihong ;
Zhou, Sanping ;
Wang, Jinjun ;
Zheng, Nanning .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :1335-1344
[9]  
Chollet F., 2015, Keras
[10]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848