Mutually-aware feature learning for few-shot object counting

被引:1
作者
Jeon, Yerim [1 ]
Lee, Subeen [1 ]
Kim, Jihwan [1 ]
Heo, Jae-Pil [1 ,2 ]
机构
[1] Sungkyunkwan Univ, Dept Artificial Intelligence, Suwon 16419, South Korea
[2] Sungkyunkwan Univ, Dept Comp Sci & Engn, Suwon 16419, South Korea
关键词
Few-shot object counting; Class-agnostic counting; Object counting; Few-shot learning; Deep learning;
D O I
10.1016/j.patcog.2024.111276
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Few-shot object counting has garnered significant attention for its practicality as it aims to count target objects in a query image based on given exemplars without additional training. However, the prevailing extract-and- match approach has a shortcoming: query and exemplar features lack interaction during feature extraction since they are extracted independently and later correlated based on similarity. This can lead to insufficient target awareness and confusion in identifying the actual target when multiple class objects coexist. To address this, we propose a novel framework, Mutually-Aware FEAture learning (MAFEA), which encodes query and exemplar features with mutual awareness from the outset. By encouraging interaction throughout the pipeline, we obtain target-aware features robust to a multi-category scenario. Furthermore, we introduce background token to effectively associate the query's target region with exemplars and decouple its background region. Our extensive experiments demonstrate that our model achieves state-of-the-art performance on FSCD-LVIS and FSC-147 benchmarks with remarkably reduced target confusion.
引用
收藏
页数:10
相关论文
共 30 条
[1]   Counting in the Wild [J].
Arteta, Carlos ;
Lempitsky, Victor ;
Zisserman, Andrew .
COMPUTER VISION - ECCV 2016, PT VII, 2016, 9911 :483-498
[2]  
Chen Y., 2024, Learning Self-Target Knowledge for Few-Shot Segmentation, V149
[3]   A Low-Shot Object Counting Network With Iterative Prototype Adaptation [J].
Dukic, Nikola ;
Lukezic, Alan ;
Zavrtanik, Vitjan ;
Kristan, Matej .
2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, :18826-18835
[4]  
Gao B.-B., 2024, CSTrans: Correlation-guided Self-Activation transformer for counting everything
[5]   Class-Agnostic Object Counting Robust to Intraclass Diversity [J].
Gong, Shenjian ;
Zhang, Shanshan ;
Yang, Jian ;
Dai, Dengxin ;
Schiele, Bernt .
COMPUTER VISION - ECCV 2022, PT XXXIII, 2022, 13693 :388-403
[6]   Masked Autoencoders Are Scalable Vision Learners [J].
He, Kaiming ;
Chen, Xinlei ;
Xie, Saining ;
Li, Yanghao ;
Dollar, Piotr ;
Girshick, Ross .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, :15979-15988
[7]   Drone-based Object Counting by Spatially Regularized Regional Proposal Network [J].
Hsieh, Meng-Ru ;
Lin, Yen-Liang ;
Hsu, Winston H. .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :4165-4173
[8]   Point, Segment and Count: A Generalized Framework for Object Counting [J].
Huang, Zhizhong ;
Dai, Mingliang ;
Zhang, Yi ;
Zhang, Junping ;
Shan, Hongming .
2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, :17067-17076
[9]   Segment Anything [J].
Kirillov, Alexander ;
Mintun, Eric ;
Ravi, Nikhila ;
Mao, Hanzi ;
Rolland, Chloe ;
Gustafson, Laura ;
Xiao, Tete ;
Whitehead, Spencer ;
Berg, Alexander C. ;
Lo, Wan-Yen ;
Dolla'r, Piotr ;
Girshick, Ross .
2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, :3992-4003
[10]  
Leibe B, 2005, PROC CVPR IEEE, P878