Diverse part attentive network for video-based person re-identification *

被引:9
作者
Shu, Xiujun [1 ,2 ]
Li, Ge [2 ]
Wei, Longhui [3 ]
Zhong, Jia-Xing [2 ]
Zang, Xianghao [2 ]
Zhang, Shiliang [4 ]
Wang, Yaowei [1 ]
Liang, Yongsheng [5 ]
Tian, Qi [1 ]
机构
[1] Peng Cheng Lab, Shenzhen 518066, Peoples R China
[2] Peking Univ, Sch Elect & Comp Engn, Shenzhen 518055, Peoples R China
[3] Sch Univ Sci & Technol China, Hefei 230026, Peoples R China
[4] Peking Univ, Sch Elect Engn & Comp Sci, Beijing 100871, Peoples R China
[5] Harbin Inst Technol, Shenzhen 518055, Peoples R China
关键词
Person re-identification; Person retrieval; Self-attention; FUSION;
D O I
10.1016/j.patrec.2021.05.020
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Attention mechanisms have achieved success in video-based person re-identification (re-ID). However, current global attentions tend to focus on the most salient parts, e.g., clothes, and ignore other subtle but valuable cues, e.g., hair, bag, and shoes. They still do not make full use of valuable information from diverse parts of human bodies. To tackle this issue, we propose a Diverse Part Attentive Network (DPAN) to exploit discriminative and diverse body cues. The framework consists of two modules: spatial diverse part attention and temporal diverse part attention. The spatial module utilizes channel grouping to exploit diverse parts of human bodies including salient and subtle parts. The temporal module aims to learn diverse weights for fusing learned features. Besides, this framework is lightweight, which introduces marginal parameters and computational complexities. Extensive experiments were conducted on three popular benchmarks, i.e. iLIDS-VID, PRID2011 and MARS. Our method achieves competitive performance on these datasets compared with state-of-the-art methods. (c) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页码:17 / 23
页数:7
相关论文
共 37 条
[1]   Video Person Re-identification with Competitive Snippet-similarity Aggregation and Co-attentive Snippet Embedding [J].
Chen, Dapeng ;
Li, Hongsheng ;
Xiao, Tong ;
Yi, Shuai ;
Wang, Xiaogang .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :CP1-CP99
[2]   Temporal Coherence or Temporal Motion: Which Is More Critical for Video-Based Person Re-identification? [J].
Chen, Guangyi ;
Rao, Yongming ;
Lu, Jiwen ;
Zhou, Jie .
COMPUTER VISION - ECCV 2020, PT VIII, 2020, 12353 :660-676
[3]   Learning Recurrent 3D Attention for Video-Based Person Re-Identification [J].
Chen, Guangyi ;
Lu, Jiwen ;
Yang, Ming ;
Zhou, Jie .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 :6963-6976
[4]   Spatial-Temporal Attention-Aware Learning for Video-Based Person Re-Identification [J].
Chen, Guangyi ;
Lu, Jiwen ;
Yang, Ming ;
Zhou, Jie .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2019, 28 (09) :4192-4205
[5]   ABD-Net: Attentive but Diverse Person Re-Identification [J].
Chen, Tianlong ;
Ding, Shaojin ;
Xie, Jingyi ;
Yuan, Ye ;
Chen, Wuyang ;
Yang, Yang ;
Ren, Zhou ;
Wang, Zhangyang .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :8350-8360
[6]  
Chen Z, 2020, P IEEE CVF C COMP VI P IEEE CVF C COMP VI
[7]  
Fu Y, 2019, AAAI CONF ARTIF INTE, P8287
[8]  
Hirzer M, 2011, LECT NOTES COMPUT SC, V6688, P91, DOI 10.1007/978-3-642-21227-7_9
[9]   VRSTC: Occlusion-Free Video Person Re-Identification [J].
Hou, Ruibing ;
Ma, Bingpeng ;
Chang, Hong ;
Gu, Xinqian ;
Shan, Shiguang ;
Chen, Xilin .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :7176-7185
[10]  
Jiang XY, 2020, AAAI CONF ARTIF INTE, V34, P11133