Relation3DMOT: Exploiting Deep Affinity for 3D Multi-Object Tracking from View Aggregation

被引:2
|
作者
Chen, Can [1 ]
Zanotti Fragonara, Luca [1 ]
Tsourdos, Antonios [1 ]
机构
[1] Cranfield Univ, Sch Aerosp Transport & Mfg, Bedford MK43 0AL, England
关键词
3D multi-object tracking; sensor fusion; deep affinity; relation learning; neural network;
D O I
10.3390/s21062113
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
Autonomous systems need to localize and track surrounding objects in 3D space for safe motion planning. As a result, 3D multi-object tracking (MOT) plays a vital role in autonomous navigation. Most MOT methods use a tracking-by-detection pipeline, which includes both the object detection and data association tasks. However, many approaches detect objects in 2D RGB sequences for tracking, which lacks reliability when localizing objects in 3D space. Furthermore, it is still challenging to learn discriminative features for temporally consistent detection in different frames, and the affinity matrix is typically learned from independent object features without considering the feature interaction between detected objects in the different frames. To settle these problems, we first employ a joint feature extractor to fuse the appearance feature and the motion feature captured from 2D RGB images and 3D point clouds, and then we propose a novel convolutional operation, named RelationConv, to better exploit the correlation between each pair of objects in the adjacent frames and learn a deep affinity matrix for further data association. We finally provide extensive evaluation to reveal that our proposed model achieves state-of-the-art performance on the KITTI tracking benchmark.
引用
收藏
页码:1 / 16
页数:16
相关论文
共 50 条
  • [21] Multi-view SoftPool attention convolutional networks for 3D model classification
    Wang, Wenju
    Wang, Xiaolin
    Chen, Gang
    Zhou, Haoran
    FRONTIERS IN NEUROROBOTICS, 2022, 16
  • [22] A multi-code 3D measurement technique based on deep learning
    Yao, Pengcheng
    Gai, Shaoyan
    Chen, Yuchong
    Chen, Wenlong
    Da, Feipeng
    OPTICS AND LASERS IN ENGINEERING, 2021, 143
  • [23] LIGHTWEIGHT MULTI-VIEW-GROUP NEURAL NETWORK FOR 3D SHAPE CLASSIFICATION
    Sun, Jiaqi
    Niu, Dongmei
    Lv, Na
    Dou, Wentao
    Peng, Jingliang
    2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 3409 - 3413
  • [24] Voxel-based three-view hybrid parallel network for 3D object classification
    Cai, Weiwei
    Liu, Dong
    Ning, Xin
    Wang, Chen
    Xie, Guojie
    DISPLAYS, 2021, 69 (69)
  • [25] Multi-Modal Fusion Based on Depth Adaptive Mechanism for 3D Object Detection
    Liu, Zhanwen
    Cheng, Juanru
    Fan, Jin
    Lin, Shan
    Wang, Yang
    Zhao, Xiangmo
    IEEE TRANSACTIONS ON MULTIMEDIA, 2025, 27 : 707 - 717
  • [26] Dual-view 3D object recognition and detection via Lidar point cloud and camera image
    Li, Jing
    Li, Rui
    Li, Jiehao
    Wang, Junzheng
    Wu, Qingbin
    Liu, Xu
    ROBOTICS AND AUTONOMOUS SYSTEMS, 2022, 150
  • [27] Multi-modal Data Analysis and Fusion for Robust Object Detection in 2D/3D Sensing
    Schierl, Jonathan
    Graehling, Quinn
    Aspiras, Theus
    Asari, Vijay
    Van Rynbach, Andre
    Rabb, Dave
    2020 IEEE APPLIED IMAGERY PATTERN RECOGNITION WORKSHOP (AIPR): TRUSTED COMPUTING, PRIVACY, AND SECURING MULTIMEDIA, 2020,
  • [28] Fast All-day 3D Object Detection Based on Multi-sensor Fusion
    Xiao, Liang
    Zhu, Qi
    Chen, Tongtong
    Zhao, Dawei
    Shang, Erke
    Nie, Yiming
    2023 IEEE CONFERENCE ON ARTIFICIAL INTELLIGENCE, CAI, 2023, : 71 - 73
  • [29] AirPose: Multi-View Fusion Network for Aeria 3D Human Pose and Shape Estimation
    Saini, Nitin
    Bonetto, Elia
    Price, Eric
    Ahmad, Aamir
    Black, Michael J.
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (02) : 4805 - 4812
  • [30] DEEP SENSOR FUSION BASED ON FRUSTUM POINT SINGLE SHOT MULTIBOX DETECTOR FOR 3D OBJECT DETECTION
    Wang, Yu
    Zhang, Ye
    Zhai, Shaohua
    Chen, Hao
    Shi, Shaoqi
    Wang, Gang
    2021 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2021, : 674 - 678