Relation3DMOT: Exploiting Deep Affinity for 3D Multi-Object Tracking from View Aggregation

被引:2
|
作者
Chen, Can [1 ]
Zanotti Fragonara, Luca [1 ]
Tsourdos, Antonios [1 ]
机构
[1] Cranfield Univ, Sch Aerosp Transport & Mfg, Bedford MK43 0AL, England
关键词
3D multi-object tracking; sensor fusion; deep affinity; relation learning; neural network;
D O I
10.3390/s21062113
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
Autonomous systems need to localize and track surrounding objects in 3D space for safe motion planning. As a result, 3D multi-object tracking (MOT) plays a vital role in autonomous navigation. Most MOT methods use a tracking-by-detection pipeline, which includes both the object detection and data association tasks. However, many approaches detect objects in 2D RGB sequences for tracking, which lacks reliability when localizing objects in 3D space. Furthermore, it is still challenging to learn discriminative features for temporally consistent detection in different frames, and the affinity matrix is typically learned from independent object features without considering the feature interaction between detected objects in the different frames. To settle these problems, we first employ a joint feature extractor to fuse the appearance feature and the motion feature captured from 2D RGB images and 3D point clouds, and then we propose a novel convolutional operation, named RelationConv, to better exploit the correlation between each pair of objects in the adjacent frames and learn a deep affinity matrix for further data association. We finally provide extensive evaluation to reveal that our proposed model achieves state-of-the-art performance on the KITTI tracking benchmark.
引用
收藏
页码:1 / 16
页数:16
相关论文
共 50 条
  • [41] Principal views selection based on growing graph convolution network for multi-view 3D model recognition
    Liang, Qi
    Li, Qiang
    Nie, Weizhi
    Su, Yuting
    APPLIED INTELLIGENCE, 2023, 53 (05) : 5320 - 5336
  • [42] Robust 3D Shape Classification Method using Simulated Multi View Sonar Images and Convolutional Nueral Network
    Lee, Meungsuk
    Kim, Jason
    Yu, Son-Cheol
    OCEANS 2019 - MARSEILLE, 2019,
  • [43] A novel multi-model 3D object detection framework with adaptive voxel-image feature fusion
    Liu, Zhao
    Fu, Zhongliang
    Li, Gang
    Zhang, Shengyuan
    IET COMPUTER VISION, 2024, 18 (05) : 640 - 651
  • [44] MV-LFN: Multi-view based local information fusion network for 3D shape recognition
    Zhang, Jing
    Zhou, Dangdang
    Zhao, Yue
    Nie, Weizhi
    Su, Yuting
    VISUAL INFORMATICS, 2021, 5 (03) : 114 - 119
  • [45] 3D Object Detection and Tracking Based on Lidar-Camera Fusion and IMM-UKF Algorithm Towards Highway Driving
    Nie, Chang
    Ju, Zhiyang
    Sun, Zhifeng
    Zhang, Hui
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2023, 7 (04): : 1242 - 1252
  • [46] 3D scene reconstruction from multi-sensor EO-SAR data
    Aksu, Ridvan
    Rahman, M. Mahbubur
    Gurbuz, Sevgi Z.
    ALGORITHMS FOR SYNTHETIC APERTURE RADAR IMAGERY XXVII, 2020, 11393
  • [47] Adaptive Multi-Pedestrian Tracking by Multi-Sensor: Track-to-Track Fusion Using Monocular 3D Detection and MMW Radar
    Zhu, Yipeng
    Wang, Tao
    Zhu, Shiqiang
    REMOTE SENSING, 2022, 14 (08)
  • [48] Markerless 3D Skeleton Tracking Algorithm by Merging Multiple Inaccurate Skeleton Data from Multiple RGB-D Sensors
    Lee, Sang-hyub
    Lee, Deok-Won
    Jun, Kooksung
    Lee, Wonjun
    Kim, Mun Sang
    SENSORS, 2022, 22 (09)
  • [49] ARC-BEV: Attentive Radar-Camera Fusion 3D Object Detection in Bird-Eye-View Space for Autonomous Driving
    Shen, Lyuyu
    Li, Jianghao
    Lee, Christina Dao Wen
    Lee, Min Young
    Hartmannsgruber, Andreas
    Ang, Marcelo H., Jr.
    EXPERIMENTAL ROBOTICS, ISER 2023, 2024, 30 : 557 - 566
  • [50] The Deep 3D Convolutional Multi-Branching Spatial-Temporal-Based Unit Predicting Citywide Traffic Flow
    Ul Abideen, Zain
    Sun, Heli
    Yang, Zhou
    Ali, Amir
    APPLIED SCIENCES-BASEL, 2020, 10 (21): : 1 - 25