Poly-MOT: A Polyhedral Framework For 3D Multi-Object Tracking

被引:20
作者
Li, Xiaoyu [1 ]
Xie, Tao [1 ]
Liu, Dedong [1 ]
Gao, Jinghan [1 ]
Dai, Kun [1 ]
Jiang, Zhiqiang [1 ]
Zhao, Lijun [1 ]
Wang, Ke [1 ]
机构
[1] Harbin Inst Technol, State Key Lab Robot & Syst, Harbin 150006, Peoples R China
来源
2023 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS) | 2023年
基金
中国国家自然科学基金;
关键词
D O I
10.1109/IROS55552.2023.10341778
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
3D Multi-object tracking (MOT) empowers mobile robots to accomplish well-informed motion planning and navigation tasks by providing motion trajectories of surrounding objects. However, existing 3D MOT methods typically employ a single similarity metric and physical model to perform data association and state estimation for all objects. With large-scale modern datasets and real scenes, there are a variety of object categories that commonly exhibit distinctive geometric properties and motion patterns. In this way, such distinctions would enable various object categories to behave differently under the same standard, resulting in erroneous matches between trajectories and detections, and jeopardizing the reliability of downstream tasks (navigation, etc.). Towards this end, we propose Poly-MOT, an efficient 3D MOT method based on the Tracking-By-Detection framework that enables the tracker to choose the most appropriate tracking criteria for each object category. Specifically, Poly-MOT leverages different motion models for various object categories to characterize distinct types of motion accurately. We also introduce the constraint of the rigid structure of objects into a specific motion model to accurately describe the highly nonlinear motion of the object. Additionally, we introduce a two-stage data association strategy to ensure that objects can find the optimal similarity metric from three custom metrics for their categories and reduce missing matches. On the NuScenes dataset, our proposed method achieves state-of-the-art performance with 75.4% AMOTA. The code is available at https://github.com/lixiaoyu2000/Poly-MOT.
引用
收藏
页码:9391 / 9398
页数:8
相关论文
共 27 条
[1]   TransFusion: Robust LiDAR-Camera Fusion for 3D Object Detection with Transformers [J].
Bai, Xuyang ;
Hu, Zeyu ;
Zhu, Xinge ;
Huang, Qingqiu ;
Chen, Yilun ;
Fu, Hangbo ;
Tai, Chiew-Lan .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, :1080-1089
[2]   Score refinement for confidence-based 3D multi-object tracking [J].
Benbarka, Nuri ;
Schroder, Jona ;
Zell, Andreas .
2021 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2021, :8083-8090
[3]  
Bewley A, 2016, IEEE IMAGE PROC, P3464, DOI 10.1109/ICIP.2016.7533003
[4]   nuScenes: A multimodal dataset for autonomous driving [J].
Caesar, Holger ;
Bankiti, Varun ;
Lang, Alex H. ;
Vora, Sourabh ;
Liong, Venice Erin ;
Xu, Qiang ;
Krishnan, Anush ;
Pan, Yu ;
Baldan, Giancarlo ;
Beijbom, Oscar .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :11618-11628
[5]   Focal Sparse Convolutional Networks for 3D Object Detection [J].
Chen, Yukang ;
Li, Yanwei ;
Zhang, Xiangyu ;
Sun, Jian ;
Jia, Jiaya .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, :5418-5427
[6]  
Chen Yukang, 2022, ARXIV220610555
[7]  
Geiger Andreas, 2012, IEEE C COMPUTER VISI, P3354
[8]  
Gwak J. Y., 2022, ARXIV E PRINTS
[9]   Joint Multi-Object Detection and Tracking with Camera-LiDAR Fusion for Autonomous Driving [J].
Huang, Kemiao ;
Hao, Qi .
2021 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2021, :6983-6989
[10]   EagerMOT: 3D Multi-Object Tracking via Sensor Fusion [J].
Kim, Aleksandr ;
Osep, Aljosa ;
Leal-Taixe, Laura .
2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, :11315-11321