Joint Spatial-Temporal Optimization for Stereo 3D Object Tracking

被引:14
|
作者
Li, Peiliang [1 ]
Shi, Jieqi [1 ]
Shen, Shaojie [1 ]
机构
[1] Hong Kong Univ Sci & Technol, Hong Kong, Peoples R China
关键词
D O I
10.1109/CVPR42600.2020.00691
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Directly learning multiple 3D objects motion from sequential images is difficult, while the geometric bundle adjustment lacks the ability to localize the invisible object centroid. To benefit from both the powerful object understanding skill from deep neural network meanwhile tackle precise geometry modeling for consistent trajectory estimation, we propose a joint spatial-temporal optimization-based stereo 3D object tracking method. From the network, we detect corresponding 2D bounding boxes on adjacent images and regress an initial 3D bounding box. Dense object cues (local depth and local coordinates) that associating to the object centroid are then predicted using a region-based network. Considering both the instant localization accuracy and motion consistency, our optimization models the relations between the object centroid and observed cues into a joint spatial-temporal error function. All historic cues will be summarized to contribute to the current estimation by a per-frame marginalization strategy without repeated computation. Quantitative evaluation on the KITTI tracking dataset shows our approach outperforms previous image-based 3D tracking methods by significant margins. We also report extensive results on multiple categories and larger datasets (KITTI raw and Argoverse Racking) for future benchmarking.
引用
收藏
页码:6876 / 6885
页数:10
相关论文
共 50 条
  • [1] 3D Single-Object Tracking with Spatial-Temporal Data Association
    Zhang, Yongchang
    Niu, Hanbing
    Guo, Yue
    He, Wenhao
    2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2022, : 264 - 269
  • [2] A spatial-temporal contexts network for object tracking
    Huang, Kai
    Xiao, Kai
    Chu, Jun
    Leng, Lu
    Dong, Xingbo
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 127
  • [3] Weakly Supervised Monocular 3D Object Detection by Spatial-Temporal View Consistency
    Han, Wencheng
    Tao, Runzhou
    Ling, Haibin
    Shen, Jianbing
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2025, 47 (01) : 84 - 98
  • [4] Object Tracking via Spatial-Temporal Memory Network
    Zhou, Zikun
    Li, Xin
    Zhang, Tianzhu
    Wang, Hongpeng
    He, Zhenyu
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (05) : 2976 - 2989
  • [5] 3D object tracking on active stereo vision robot
    Shibata, M
    Honma, T
    7TH INTERNATIONAL WORKSHOP ON ADVANCED MOTION CONTROL, PROCEEDINGS, 2002, : 567 - 572
  • [6] You Don't Only Look Once: Constructing Spatial-Temporal Memory for Integrated 3D Object Detection and Tracking
    Sun, Jiaming
    Xie, Yiming
    Zhang, Siyu
    Chen, Linghao
    Zhang, Guofeng
    Bao, Hujun
    Zhou, Xiaowei
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 3165 - 3174
  • [7] Fast spatial-temporal stereo matching for 3D face reconstruction under speckle pattern projection
    Fu, Keren
    Xie, Yijiang
    Jing, Hailong
    Zhu, Jiangping
    IMAGE AND VISION COMPUTING, 2019, 85 : 36 - 45
  • [8] Spatial-temporal correlations in the speckle pattern for the characterization of cellular motion within a 3D object
    Weil, Yonni
    Shafran, Yana
    Sobolev, Maria
    Afrimzon, Lena
    Zurgil, Naomi
    Deutsch, Motti
    Schiffer, Zeev
    BIOMEDICAL OPTICS EXPRESS, 2023, 14 (05) : 1974 - 1991
  • [9] Online learning and joint optimization of combined spatial-temporal models for robust visual tracking
    Zhou, Tao
    Bhaskar, Harish
    Liu, Fanghui
    Yang, Jie
    Cai, Ping
    NEUROCOMPUTING, 2017, 226 : 221 - 237
  • [10] An improved spatial-temporal regularization method for visual object tracking
    Hayat, Muhammad Umar
    Ali, Ahmad
    Khan, Baber
    Mehmood, Khizer
    Ullah, Khitab
    Amir, Muhammad
    SIGNAL IMAGE AND VIDEO PROCESSING, 2024, 18 (03) : 2065 - 2077