Cascaded attention-guided multi-granularity feature learning for person re-identification

被引:8
作者
Dong, Husheng [1 ,2 ]
Yang, Yuanfeng [1 ]
Sun, Xun [1 ]
Zhang, Liang [1 ]
Fang, Ligang [1 ]
机构
[1] Suzhou Vocat Univ, Sch Comp Engn, Suzhou 215104, Peoples R China
[2] Jiangsu Prov Support Software Engn R&D Ctr Modern, Suzhou 215104, Peoples R China
关键词
Person re-identification; Attention; Feature representation; Neural network; Visual clues; NETWORK;
D O I
10.1007/s00138-022-01353-3
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Attention mechanism has been extensively employed in the task of person re-identification, as it helps to extract much more discriminative feature representations. However, most of existing works either incorporate a single-scale attention module, or the embedded attentions work independently. Though promising results are achieved, they may fail to mine different subtle visual clues. To mitigate this issue, a novel framework called cascaded attention network (CANet) is proposed, which allows to mine diverse clues and integrate them into final multi-granularity features by a cascaded manner. Specifically, we design a novel hybrid pooling attention module (HPAM) and plug it into backbone network at different stages. To make them work collaboratively, an inter-attention regularization is applied, such that they can localize complementary salient features. Then, CANet extracts global and local features from a part-based pyramidal architecture. For better feature robustness, supervision is applied to not only the pyramidal branches, but also those intermediate attention modules. Furthermore, within each supervision branch, hybrid pooling with two different strides is executed to enhance feature representation capabilities. Extensive experiments with ablation analysis demonstrate the effectiveness of the proposed method, and state-of-the-art results are achieved on three public benchmark datasets, including Market-1501, CUHK03, and DukeMTMC-ReID.
引用
收藏
页数:16
相关论文
共 62 条
[1]  
Ahmed E, 2015, PROC CVPR IEEE, P3908, DOI 10.1109/CVPR.2015.7299016
[2]  
[Anonymous], 2015, P IEEE C COMP VIS PA
[3]   Mixed High-Order Attention Network for Person Re-Identification [J].
Chen, Binghui ;
Deng, Weihong ;
Hu, Jiani .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :371-381
[4]   ICE: Inter-instance Contrastive Encoding for Unsupervised Person Re-identification [J].
Chen, Hao ;
Lagadec, Benoit ;
Bremond, Francois .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :14940-14949
[5]   Occlude Them All: Occlusion-Aware Attention Network for Occluded Person Re-ID [J].
Chen, Peixian ;
Liu, Wenfeng ;
Dai, Pingyang ;
Liu, Jianzhuang ;
Ye, Qixiang ;
Xu, Mingliang ;
Chen, Qi'an ;
Ji, Rongrong .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :11813-11822
[6]   ABD-Net: Attentive but Diverse Person Re-Identification [J].
Chen, Tianlong ;
Ding, Shaojin ;
Xie, Jingyi ;
Yuan, Ye ;
Chen, Wuyang ;
Yang, Yang ;
Ren, Zhou ;
Wang, Zhangyang .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :8350-8360
[7]   Beyond triplet loss: a deep quadruplet network for person re-identification [J].
Chen, Weihua ;
Chen, Xiaotang ;
Zhang, Jianguo ;
Huang, Kaiqi .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :1320-1329
[8]   Salience-Guided Cascaded Suppression Network for Person Re-identification [J].
Chen, Xuesong ;
Fu, Canmiao ;
Zhao, Yong ;
Zheng, Feng ;
Song, Jingkuan ;
Ji, Rongrong ;
Yang, Yi .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, :3297-3307
[9]   Deep attention aware feature learning for person re-Identification [J].
Chen, Yifan ;
Wang, Han ;
Sun, Xiaolu ;
Fan, Bin ;
Tang, Chu ;
Zeng, Hui .
PATTERN RECOGNITION, 2022, 126
[10]   Person Re-Identification by Multi-Channel Parts-Based CNN with Improved Triplet Loss Function [J].
Cheng, De ;
Gong, Yihong ;
Zhou, Sanping ;
Wang, Jinjun ;
Zheng, Nanning .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :1335-1344