Relation3DMOT: Exploiting Deep Affinity for 3D Multi-Object Tracking from View Aggregation

被引:2
|
作者
Chen, Can [1 ]
Zanotti Fragonara, Luca [1 ]
Tsourdos, Antonios [1 ]
机构
[1] Cranfield Univ, Sch Aerosp Transport & Mfg, Bedford MK43 0AL, England
关键词
3D multi-object tracking; sensor fusion; deep affinity; relation learning; neural network;
D O I
10.3390/s21062113
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
Autonomous systems need to localize and track surrounding objects in 3D space for safe motion planning. As a result, 3D multi-object tracking (MOT) plays a vital role in autonomous navigation. Most MOT methods use a tracking-by-detection pipeline, which includes both the object detection and data association tasks. However, many approaches detect objects in 2D RGB sequences for tracking, which lacks reliability when localizing objects in 3D space. Furthermore, it is still challenging to learn discriminative features for temporally consistent detection in different frames, and the affinity matrix is typically learned from independent object features without considering the feature interaction between detected objects in the different frames. To settle these problems, we first employ a joint feature extractor to fuse the appearance feature and the motion feature captured from 2D RGB images and 3D point clouds, and then we propose a novel convolutional operation, named RelationConv, to better exploit the correlation between each pair of objects in the adjacent frames and learn a deep affinity matrix for further data association. We finally provide extensive evaluation to reveal that our proposed model achieves state-of-the-art performance on the KITTI tracking benchmark.
引用
收藏
页码:1 / 16
页数:16
相关论文
共 50 条
  • [1] Exploiting Multi-Modal Synergies for Enhancing 3D Multi-Object Tracking
    Xu, Xinglong
    Ren, Weihong
    Chen, Xi'ai
    Fan, Huijie
    Han, Zhi
    Liu, Honghai
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (10): : 8643 - 8650
  • [2] Deformable Feature Aggregation for Dynamic Multi-modal 3D Object Detection
    Chen, Zehui
    Li, Zhenyu
    Zhang, Shiquan
    Fang, Liangji
    Jiang, Qinhong
    Zhao, Feng
    COMPUTER VISION, ECCV 2022, PT VIII, 2022, 13668 : 628 - 644
  • [3] Manipulator and object tracking for in-hand 3D object modeling
    Krainin, Michael
    Henry, Peter
    Ren, Xiaofeng
    Fox, Dieter
    INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2011, 30 (11) : 1311 - 1327
  • [4] CFMVOR: Federated Multi-view 3D Object Recognition Based on Compressed Learning
    Xiao, Di
    Zhang, Meng
    Zhang, Maolan
    Chen, Lvjun
    PATTERN RECOGNITION AND COMPUTER VISION, PT XIII, PRCV 2024, 2025, 15043 : 280 - 293
  • [5] Practical 3D human skeleton tracking based on multi-view and multi-Kinect fusion
    Nguyen, Manh-Hung
    Hsiao, Ching-Chun
    Cheng, Wen-Huang
    Huang, Ching-Chun
    MULTIMEDIA SYSTEMS, 2022, 28 (02) : 529 - 552
  • [6] Practical 3D human skeleton tracking based on multi-view and multi-Kinect fusion
    Manh-Hung Nguyen
    Ching-Chun Hsiao
    Wen-Huang Cheng
    Ching-Chun Huang
    Multimedia Systems, 2022, 28 : 529 - 552
  • [7] Deep Learning Based 3D Object Detection for Automotive Radar and Camera
    Meyer, Michael
    Kuschk, Georg
    2019 16TH EUROPEAN RADAR CONFERENCE (EURAD), 2019, : 133 - 136
  • [8] RoIFusion: 3D Object Detection From LiDAR and Vision
    Chen, Can
    Fragonara, Luca Zanotti
    Tsourdos, Antonios
    IEEE ACCESS, 2021, 9 (09): : 51710 - 51721
  • [9] Multi-Modal Streaming 3D Object Detection
    Abdelfattah, Mazen
    Yuan, Kaiwen
    Wang, Z. Jane
    Ward, Rabab
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2023, 8 (10) : 6163 - 6170
  • [10] Online Multi-object Tracking from A Bird's-eye View by Fusion of Millimeter-wave Radar and Vision
    Zhang, Qiang
    Song, Yuying
    Li, Zecheng
    Ai, Fuyuan
    Song, Chunyi
    Xu, Zhiwei
    2024 IEEE RADAR CONFERENCE, RADARCONF 2024, 2024,