Multi-scale joint learning for person re-identification

被引:0
|
作者
Xie P. [1 ]
Xu X. [1 ,2 ,3 ]
机构
[1] School of Computer Science and Technology, Wuhan University of Science and Technology, Wuhan
[2] Hubei Province Key Laboratory of Intelligent Information Processing and Real-time Industrial System, Wuhan University of Science and Technology, Wuhan
[3] School of Electronic Information and Electrical Engineering, Shanghai Jiao Tong University, Shanghai
来源
Beijing Hangkong Hangtian Daxue Xuebao/Journal of Beijing University of Aeronautics and Astronautics | 2021年 / 47卷 / 03期
基金
中国国家自然科学基金;
关键词
Deep learning; Joint learning; Multi-branch network; Multi-scale; Person re-identification;
D O I
10.13700/j.bh.1001-5965.2020.0445
中图分类号
学科分类号
摘要
The existing person re-identification approaches mainly focus on learning person's local features to match a specific pedestrian across different cameras. However, in the presence of incomplete conditions of pedestrian data such as motion or occlusion of human body parts, background interference, etc., it leads to an increase in the probability of partial loss of pedestrian recognition information. This paper presents a multi-scale joint learning method to extract the fine-grained person feature. This method consists of three subnets, i. e. coarse-grained global feature extraction subnet, fine-grained global feature extraction subnet, and fine-grained local feature extraction subnet. The coarse-grained global feature extraction subnet enhances the diversity of the global feature by fusing semantic information at different levels. The fine-grained global branching unites all local features to learn the correlation among local components of a pedestrian while describing the global features at a fine-grained level. The fine-grained local feature extraction subnet enhances robustness by traversing local features and finding out pedestrian non-significant information. Comparative experiments have been conducted to evaluate the performance of the proposed method against state-of-the-art methods on Market1501, DukeMTMC-ReID, and CUHK03 person re-identification datasets. The experimental results show that the proposed method has the best performance. © 2021, Editorial Board of JBUAA. All right reserved.
引用
收藏
页码:613 / 622
页数:9
相关论文
共 40 条
  • [1] LECUN Y, BENGIO Y, HINTON G., Deep learning, Nature, 521, 7553, pp. 436-444, (2015)
  • [2] ZHAO H, TIAN M, SUN S, Et al., Spindle Net: Person re-identification with human body region guided feature decomposition and fusion, Proceedings of the IEEE International Conference on Computer Vision, pp. 907-915, (2017)
  • [3] SUN Y, ZHENG L, YANG Y, Et al., Beyond part models: Person retrieval with refined part pooling (and a strong convolutional baseline), Proceedings of the European Conference on Computer Vision, pp. 480-496, (2018)
  • [4] ZHENG L, HUANG Y, LU H, Et al., Pose-invariant embedding for deep person re-identification, IEEE Transactions on Image Processing, 28, 9, pp. 4500-4509, (2019)
  • [5] WEI L, ZHANG S, YAO H, Et al., GLAD: Global-local-alignment descriptor for pedestrian retrieval, Proceedings of the 25th ACM International Conference on Multimedia, pp. 420-428, (2017)
  • [6] ZHENG F, DENG C, SUN X, Et al., Pyramidal person re-identification via multi-loss dynamic training, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8514-8522, (2019)
  • [7] FU Y, WEI Y, ZHOU Y, Et al., Horizontal pyramid matching for person re-identification, Proceedings of the AAAI Conference on Artificial Intelligence, pp. 8295-8302, (2019)
  • [8] WANG G, YUAN Y, CHEN X, Et al., Learning discriminative features with multiple granularities for person re-identification, Proceedings of the 26th ACM International Conference on Multimedia, pp. 274-282, (2018)
  • [9] WANG Z, JIANG J, WU Y, Et al., Learning sparse and identity-preserved hidden attributes for person re-identification, IEEE Transactions on Image Processing, 29, 1, pp. 2013-2025, (2019)
  • [10] ZENG Z, WANG Z, WANG Z, Et al., Illumination-adaptive person re-identification, IEEE Transactions on Multimedia, 22, 12, pp. 3064-3074, (2020)