An Effective Motion-Centric Paradigm for 3D Single Object Tracking in Point Clouds

被引:1
作者
Zheng, Chaoda [1 ,2 ]
Yan, Xu [1 ,2 ]
Zhang, Haiming [1 ,2 ]
Wang, Baoyuan [3 ]
Cheng, Shenghui [4 ]
Cui, Shuguang [1 ,2 ]
Li, Zhen [1 ,2 ]
机构
[1] Chinese Univ Hong Kong, Future Network Intelligence Inst FNii, Shenzhen 518172, Peoples R China
[2] Chinese Univ Hong Kong, Sch Sci & Engn SSE, Shenzhen 518172, Peoples R China
[3] Xiaobing AI, Beijing 100032, Peoples R China
[4] Westlake Univ, Hangzhou 310024, Zhejiang, Peoples R China
关键词
Single object tracking; point cloud; LiDAR; motion; semi-supervised learning;
D O I
10.1109/TPAMI.2023.3324372
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
3D single object tracking in LiDAR point clouds (LiDAR SOT) plays a crucial role in autonomous driving. Current approaches all follow the Siamese paradigm based on appearance matching. However, LiDAR point clouds are usually textureless and incomplete, which hinders effective appearance matching. Besides, previous methods greatly overlook the critical motion clues among targets. In this work, beyond 3D Siamese tracking, we introduce a motion-centric paradigm to handle LiDAR SOT from a new perspective. Following this paradigm, we propose a matching-free two-stage tracker M-2-Track. At the 1st-stage, M-2-Track localizes the target within successive frames via motion transformation. Then it refines the target box through motion-assisted shape completion at the 2nd-stage. Due to the motion-centric nature, our method shows its impressive generalizability with limited training labels and provides good differentiability for end-to-end cycle training. This inspires us to explore semi-supervised LiDAR SOT by incorporating a pseudo-label-based motion augmentation and a self-supervised loss term. Under the fully-supervised setting, extensive experiments confirm that M-2-Track significantly outperforms previous state-of-the-arts on three large-scale datasets while running at 57FPS (similar to 3%, similar to 11% and similar to 22% precision gains on KITTI, NuScenes, and Waymo Open Dataset respectively). While under the semi-supervised setting, our method performs on par with or even surpasses its fully-supervised counterpart using fewer than half of the labels from KITTI. Further analysis verifies each component's effectiveness and shows the motion-centric paradigm's promising potential for auto-labeling and unsupervised domain adaptation.
引用
收藏
页码:43 / 60
页数:18
相关论文
共 64 条
  • [11] Spatio-Temporal Contextual Learning for Single Object Tracking on Point Clouds
    Gao, Jiantao
    Yan, Xu
    Zhao, Weibing
    Lyu, Zhen
    Liao, Yinghong
    Zheng, Chaoda
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (07) : 9470 - 9482
  • [12] Geiger A, 2012, PROC CVPR IEEE, P3354, DOI 10.1109/CVPR.2012.6248074
  • [13] Leveraging Shape Completion for 3D Siamese Tracking
    Giancola, Silvio
    Zarzar, Jesus
    Ghanem, Bernard
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 1359 - 1368
  • [14] Hu PY, 2020, PROC CVPR IEEE, P10998, DOI 10.1109/CVPR42600.2020.01101
  • [15] Hui L, 2021, ADV NEUR IN, V34
  • [16] Guided Point Contrastive Learning for Semi-supervised Point Cloud Semantic Segmentation
    Jiang, Li
    Shi, Shaoshuai
    Tian, Zhuotao
    Lai, Xin
    Liu, Shu
    Fu, Chi-Wing
    Jia, Jiaya
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 6403 - 6412
  • [17] Deformation and Correspondence Aware Unsupervised Synthetic-to-Real Scene Flow Estimation for Point Clouds
    Jin, Zhao
    Lei, Yinjie
    Akhtar, Naveed
    Li, Haifeng
    Hayat, Munawar
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 7223 - 7233
  • [18] EagerMOT: 3D Multi-Object Tracking via Sensor Fusion
    Kim, Aleksandr
    Osep, Aljosa
    Leal-Taixe, Laura
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 11315 - 11321
  • [19] Lan Kaihao, 2022, P AS C COMP VIS, P399
  • [20] SiamRPN plus plus : Evolution of Siamese Visual Tracking with Very Deep Networks
    Li, Bo
    Wu, Wei
    Wang, Qiang
    Zhang, Fangyi
    Xing, Junliang
    Yan, Junjie
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 4277 - 4286