Few-shot learning for novel object detection in autonomous driving

被引:0
作者
Zhuang, Yifan [1 ]
Liu, Pei [2 ]
Yang, Hao [1 ]
Zhang, Kai [3 ]
Wang, Yinhai [1 ]
Pu, Ziyuan [4 ]
机构
[1] Department of Civil and Environmental Engineering, University of Washington, Seattle
[2] Intelligent Transportation Thrust, Systems Hub, The Hong Kong University of Science and Technology (Guangzhou), Guangzhou
[3] Tsinghua Shenzhen International Graduate School, Tsinghua University, Shenzhen
[4] School of Transportation, Southeast University, Nanjing
来源
Communications in Transportation Research | 2025年 / 5卷
关键词
Autonomous driving; Computer vision; Few-shot learning; Object detection;
D O I
10.1016/j.commtr.2025.100194
中图分类号
学科分类号
摘要
Artificial intelligence and advanced sensing technologies have significantly advanced the intelligent transportation system and autonomous vehicles. Perception, a critical component, extracts real-time traffic information essential for various system functionalities, such as agent behavior prediction. However, the quality of information derived from perception greatly influences overall system performance. This study focuses on enhancing perception robustness in autonomous vehicles, particularly in detecting rare objects, which pose a challenge due to limited training samples. While deep learning-based vision methods have shown promising accuracy, they struggle with rare object detection. To address this, we propose a few-shot learning training strategy tailored for improved detection accuracy of rare or novel objects. Additionally, we design a one-stage object detector for efficient object detection in autonomous driving scenarios. Experiments on a self-driving dataset augmented with rare objects alongside the popular few-shot object detection (FSOD) benchmark, the pattern analysis, statical modeling, and computational learning PASCAL Visual Object Classes (PASCAL-VOC), demonstrate state-of-the-art accuracy in rare categories and superior inference speed compared to alternative algorithms. Furthermore, we investigate the impact of intra-class variance on detection accuracy, providing insights for data annotation in the preparation stage. © 2025 The Authors
引用
收藏
相关论文
共 52 条
[1]  
Ackermann S., Schawinski K., Zhang C., Weigel A.K., Turp M.D., Using transfer learning to detect galaxy mergers, Mon. Not. Roy. Astron. Soc., 479, pp. 415-425, (2018)
[2]  
Agarwal A., Majee A., Subramanian A., Arora C., Attention guided cosine margin to overcome class-imbalance in few-shot road object detection, Proc. IEEE/CVF Winter Conf. Appl. Comput. Vis., pp. 221-230, (2022)
[3]  
Chen H., Wang Y., Wang G., Qiao Y., Lstd: a low-shot transfer detector for object detection, Proc. AAAI Conf. Artif. Intell., (2018)
[4]  
Chen Q., Wang Y., Yang T., Zhang X., Cheng J., Sun J., You only look one-level feature, Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., pp. 13039-13048, (2021)
[5]  
Dong Y., Kang C., Zhang J., Zhu Z., Wang Y., Yang X., Et al., Benchmarking robustness of 3d object detection to common corruptions, Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., pp. 1022-1032, (2023)
[6]  
Everingham M., Van Gool L., Williams C.K., Winn J., Zisserman A., The pascal visual object classes (voc) challenge, Int. J. Comput. Vis., 88, pp. 303-338, (2010)
[7]  
Fan Z., Yu J.-G., Liang Z., Ou J., Gao C., Xia G.-S., Et al., Fgn: fully guided network for few-shot instance segmentation, Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., pp. 9172-9181, (2020)
[8]  
Feng D., Harakeh A., Waslander S.L., Dietmayer K., A review and comparative study on probabilistic object detection in autonomous driving, IEEE Trans. Intell. Transport. Syst., 23, pp. 9961-9980, (2021)
[9]  
Gazzaley A., Nobre A.C., Top-down modulation: bridging selective attention and working memory, Trends Cognit. Sci., 16, pp. 129-135, (2012)
[10]  
Geiger A., Lenz P., Stiller C., Urtasun R., Vision meets robotics: the kitti dataset, Int. J. Robot Res., 32, pp. 1231-1237, (2013)