Semantically Guided Bi-level Adaptation for Cross Domain Crowd Counting

被引:0
作者
Zhao, Muming [1 ]
Xu, Weiqing [2 ]
Zhang, Chongyang [2 ]
机构
[1] Beijing Forestry Univ, Beijing, Peoples R China
[2] Shanghai Jiao Tong Univ, Sch Elect Informat & Elect Engn, Shanghai, Peoples R China
来源
PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT XI | 2024年 / 14435卷
关键词
Crowd counting; Task correspondence; Domain adaptation;
D O I
10.1007/978-981-99-8552-4_26
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Visual crowd counting has played an important role in various practical applications. However, domain gap remains a major barrier preventing models trained on the source domain (e.g., training scenes) generalize well to the target domain (e.g., unseen testing scenes). Crowd semantic information are shown to be beneficial to assist crowd counting in supervised training settings, implying the close relationship between crowd density and semantics. Nevertheless, the potential of this powerful cue has bot been fully explored in the unsupervised domain adaptation (UDA) setting. Motivated by the observation that crowd density map share domain-invariant correspondence with the crowd segmentation map, we propose to adapt this correspondence correlation from the source domain to the target domain to address the domain gap. To this end, a semantically guided task correlation layer is introduced to extract the task correspondence map, whose coherence is enforced across domains by adversarial training. To drive the adaption of earlier hidden layers directly, we further align the task correspondence correlation upon intermediate-level outputs. Extensive experiments are conducted on three benchmark datasets. The performances of our method either surpass or are on par with the counterparts, demonstrating the effectiveness of the proposed approach for cross-domain crowd counting.
引用
收藏
页码:327 / 338
页数:12
相关论文
共 28 条
[1]  
Boski M, 2017, 2017 10TH INTERNATIONAL WORKSHOP ON MULTIDIMENSIONAL (ND) SYSTEMS (NDS)
[2]   Decoupled Two-Stage Crowd Counting and Beyond [J].
Cheng, Jian ;
Xiong, Haipeng ;
Cao, Zhiguo ;
Lu, Hao .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 :2862-2875
[3]  
Du ZP, 2023, Arxiv, DOI arXiv:2212.02573
[4]  
Ganin Y, 2015, PR MACH LEARN RES, V37, P1180
[5]   Bi-level Alignment for Cross-Domain Crowd Counting [J].
Gong, Shenjian ;
Zhang, Shanshan ;
Yang, Jian ;
Dai, Dengxin ;
Schiele, Bernt .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, :7532-7540
[6]  
Han T, 2020, INT CONF ACOUST SPEE, P1848, DOI [10.1109/icassp40776.2020.9054768, 10.1109/ICASSP40776.2020.9054768]
[7]   Composition Loss for Counting, Density Map Estimation and Localization in Dense Crowds [J].
Idrees, Haroon ;
Tayyab, Muhmmad ;
Athrey, Kishan ;
Zhang, Dong ;
Al-Maadeed, Somaya ;
Rajpoot, Nasir ;
Shah, Mubarak .
COMPUTER VISION - ECCV 2018, PT II, 2018, 11206 :544-559
[8]   ImageNet Classification with Deep Convolutional Neural Networks [J].
Krizhevsky, Alex ;
Sutskever, Ilya ;
Hinton, Geoffrey E. .
COMMUNICATIONS OF THE ACM, 2017, 60 (06) :84-90
[9]   CSRNet: Dilated Convolutional Neural Networks for Understanding the Highly Congested Scenes [J].
Li, Yuhong ;
Zhang, Xiaofan ;
Chen, Deming .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :1091-1100
[10]   Crowd Counting with Deep Structured Scale Integration Network [J].
Liu, Lingbo ;
Qiu, Zhilin ;
Li, Guanbin ;
Liu, Shufan ;
Ouyang, Wanli ;
Lin, Liang .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :1774-1783