Exploring Diversity-Based Active Learning for 3D Object Detection in Autonomous Driving

被引:1
|
作者
Lin, Jinpeng [1 ]
Liang, Zhihao [2 ]
Deng, Shengheng [3 ]
Cai, Lile [4 ]
Jiang, Tao [5 ]
Li, Tianrui [1 ]
Jia, Kui [6 ]
Xu, Xun [4 ]
机构
[1] Southwest Jiaotong Univ, Sch Comp & Artificial Intelligence, Chengdu 610032, Peoples R China
[2] South China Univ Technol, Sch Elect & Informat Engn, Guangzhou 510640, Peoples R China
[3] Weride, Guangzhou 510700, Huangpu, Peoples R China
[4] ASTAR, Inst Infocomm Res I2R, Singapore 138632, Singapore
[5] Chengdu Univ Tradit Chinese Med, Sch Intelligent Med, Chengdu 610075, Peoples R China
[6] Chinese Univ Hong Kong, Sch Data Sci, Shenzhen, Peoples R China
基金
中国国家自然科学基金;
关键词
Three-dimensional displays; Object detection; Costs; Annotations; Detectors; Uncertainty; Diversity reception; Feature extraction; Autonomous vehicles; Point cloud compression; Active learning; 3D object detection; autonomous driving;
D O I
10.1109/TITS.2024.3463801
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
3D object detection has recently received much attention due to its great potential in autonomous vehicle (AV). The success of deep learning based object detectors relies on the availability of large-scale annotated datasets, which is time-consuming and expensive to compile, especially for 3D bounding box annotation. In this work, we investigate diversity-based active learning (AL) as a potential solution to alleviate the annotation burden. Given limited annotation budget, only the most informative frames and objects are automatically selected for human to annotate. Technically, we take the advantage of the multimodal information provided in an AV dataset, and propose a novel acquisition function that enforces spatial and temporal diversity in the selected samples. We benchmark the proposed method against other AL strategies under realistic annotation cost measurements, where the realistic costs for annotating a frame and a 3D bounding box are both taken into consideration. We demonstrate the effectiveness of the proposed method on the nuScenes dataset and show that it outperforms existing AL strategies significantly.
引用
收藏
页码:15454 / 15466
页数:13
相关论文
共 50 条
  • [1] Diversity-Based Active Learning: Creating a Representative Object Detection Dataset in 3D Point Clouds
    Moses, Aimee
    Bogart, Christopher W.
    O'Haire, Kelsey
    Solorzano, Lidia
    Yeo, Elinor
    GEOSPATIAL INFORMATICS XIII, 2023, 12525
  • [2] A review of 3D object detection based on autonomous driving
    Wang, Huijuan
    Chen, Xinyue
    Yuan, Quanbo
    Liu, Peng
    VISUAL COMPUTER, 2025, 41 (03): : 1757 - 1775
  • [3] Survey on deep learning-based 3D object detection in autonomous driving
    Liang, Zhenming
    Huang, Yingping
    TRANSACTIONS OF THE INSTITUTE OF MEASUREMENT AND CONTROL, 2023, 45 (04) : 761 - 776
  • [4] LiDAR-based 3D Object Detection for Autonomous Driving
    Li, Zirui
    2022 INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, COMPUTER VISION AND MACHINE LEARNING (ICICML), 2022, : 507 - 512
  • [5] Monocular 3D Object Detection for Autonomous Driving
    Chen, Xiaozhi
    Kundu, Kaustav
    Zhang, Ziyu
    Ma, Huimin
    Fidler, Sanja
    Urtasun, Raquel
    2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 2147 - 2156
  • [6] 3D Object Detection for Autonomous Driving: A Survey
    Qian, Rui
    Lai, Xin
    Li, Xirong
    PATTERN RECOGNITION, 2022, 130
  • [7] A Survey on Deep-Learning-Based LiDAR 3D Object Detection for Autonomous Driving
    Alaba, Simegnew Yihunie
    Ball, John E.
    SENSORS, 2022, 22 (24)
  • [8] 3D object detection based on image and LIDAR fusion for autonomous driving
    Chen G.
    Yi H.
    Mao Z.
    International Journal of Vehicle Information and Communication Systems, 2023, 8 (03) : 237 - 251
  • [9] R-CNN Based 3D Object Detection for Autonomous Driving
    Hu, Hongyu
    Zhao, Tongtong
    Wang, Qi
    Gao, Fei
    He, Lei
    CICTP 2020: TRANSPORTATION EVOLUTION IMPACTING FUTURE MOBILITY, 2020, : 918 - 929
  • [10] Monocular 3D Object Detection for Autonomous Driving Based on Contextual Transformer
    She, Xiangyang
    Yan, Weijia
    Dong, Lihong
    Computer Engineering and Applications, 2024, 60 (19) : 178 - 189