Temporal Point Cloud Fusion With Scene Flow for Robust 3D Object Tracking

被引:7
作者
Yang, Yanding [1 ,2 ]
Jiang, Kun [1 ]
Yang, Diange [1 ]
Jiang, Yanqin [3 ]
Lu, Xiaowei [3 ]
机构
[1] Tsinghua Univ, Sch Vehicle & Mobil, Beijing 100084, Peoples R China
[2] Dongfeng Motor Corp, Wuhan 430056, Peoples R China
[3] Natl Innovat Ctr Intelligent & Connected Vehicles, Beijing 100176, Peoples R China
基金
国家重点研发计划;
关键词
Three-dimensional displays; Feature extraction; Point cloud compression; Object tracking; Estimation; Training; Tracking; Scene flow estimation; feature fusion; 3D object tracking; NETWORK;
D O I
10.1109/LSP.2022.3185948
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Non-visual range sensors such as Lidar have shown the potential to detect, locate and track objects in complex dynamic scenes thanks to their higher stability in comparison with vision-based sensors like cameras. However, due to the disorder, sparsity, and irregularity of the point cloud, it is much more challenging to take advantage of the temporal information in the dynamic 3D point cloud sequences, as it has been done in the image sequences for improving detection and tracking. In this paper, we propose a novel scene-flow-based point cloud feature fusion module to tackle this challenge, based on which a 3D object tracking framework is also achieved to exploit the temporal motion information. Moreover, we carefully designed several training schemes that contribute to the success of this new module by eliminating the issues of overfitting and long-tailed distribution of object categories. Extensive experiments on the public KITTI 3D object tracking dataset demonstrate the effectiveness of the proposed method by achieving superior results to the baselines.
引用
收藏
页码:1579 / 1583
页数:5
相关论文
共 35 条
[1]  
Baser E, 2019, IEEE INT VEH SYM, P1426, DOI [10.1109/ivs.2019.8813779, 10.1109/IVS.2019.8813779]
[2]  
Bernardin K., 2006, 6 IEEE INT WORKSH VI, VVolume 90
[3]   A METHOD FOR REGISTRATION OF 3-D SHAPES [J].
BESL, PJ ;
MCKAY, ND .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 1992, 14 (02) :239-256
[4]   Joint Anchor-Feature Refinement for Real-Time Accurate Object Detection in Images and Videos [J].
Chen, Xingyu ;
Yu, Junzhi ;
Kong, Shihan ;
Wu, Zhengxing ;
Wen, Li .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2021, 31 (02) :594-607
[5]  
Chiu HK, 2020, Arxiv, DOI arXiv:2001.05673
[6]   Relation Distillation Networks for Video Object Detection [J].
Deng, Jiajun ;
Pan, Yingwei ;
Yao, Ting ;
Zhou, Wengang ;
Li, Houqiang ;
Mei, Tao .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :7022-7031
[7]  
Dewan A, 2016, 2016 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS 2016), P1765, DOI 10.1109/IROS.2016.7759282
[8]   The PASCAL Visual Object Classes Challenge: A Retrospective [J].
Everingham, Mark ;
Eslami, S. M. Ali ;
Van Gool, Luc ;
Williams, Christopher K. I. ;
Winn, John ;
Zisserman, Andrew .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2015, 111 (01) :98-136
[9]   Detect to Track and Track to Detect [J].
Feichtenhofer, Christoph ;
Pinz, Axel ;
Zisserman, Andrew .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :3057-3065
[10]  
Fekete L, 2005, HUNGARY