Fast Segmentation-Based Object Tracking Model for Autonomous Vehicles

被引:5
作者
Dong, Xiaoyun [1 ,2 ,3 ]
Niu, Jianwei [1 ,2 ,3 ,5 ]
Cui, Jiahe [1 ,2 ,3 ]
Fu, Zongkai [1 ,2 ,3 ]
Ouyang, Zhenchao [1 ,2 ,4 ]
机构
[1] Beihang Univ, Hangzhou Innovat Inst, Chuanghui St 18, Hangzhou 310000, Zhejiang, Peoples R China
[2] State Key Lab Virtual Real Technol & Syst, Beijing, Peoples R China
[3] Beihang Univ, Beijing Adv Innovat Ctr Big Data & Brain Comp BDB, Xueyuan Rd 37, Beijing 100191, Peoples R China
[4] Nanhu Lab, Jiaxin 314000, Zhejiang, Peoples R China
[5] Zhengzhou Univ, Zhengzhou Univ Res Inst Ind Technol, Zhengzhou 450001, Peoples R China
来源
ALGORITHMS AND ARCHITECTURES FOR PARALLEL PROCESSING, ICA3PP 2020, PT II | 2020年 / 12453卷
基金
中国国家自然科学基金;
关键词
Object tracking; Segmentation; Bi-LSTM; Autonomous vehicle; Deep learning;
D O I
10.1007/978-3-030-60239-0_18
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
On-road object tracking is a criticalmodule for both Advanced Driving Assistant System (ADAS) and autonomous vehicles. Commonly, this function can be achieved through single vehicle sensors, such as a camera or LiDAR. Consider the low cost and wide application of optical cameras, a simple image segmentation-based on-road object tracking model is proposed. Different from the detection-based tracking with bounding box, our model improves tracking performance from the following three aspects: 1) the Positional Normalization (PONO) feature is used to enhance the target outline with common convolutional layers. 2) The inter-frame correlation of each target used for tracking relies on mask, this helps the model reducing the influences caused by the background around the targets. 3) By using a bidirectional LSTM module capable of capturing timing correlation information, the forward and reverse matching of the targets in consecutive frames is performed. We also evaluate the presented model on the KITTI MOTS (Multi-Object and Segmentation) task which collected from out door environment for autonomous vehicle. Results show that our model is three times faster than Track RCNN with slightly drop on sMOTSA, and is more suitable for deployment on vehicular low-power edge computing equipment.
引用
收藏
页码:259 / 273
页数:15
相关论文
共 39 条
[1]  
Bewley A, 2016, IEEE IMAGE PROC, P3464, DOI 10.1109/ICIP.2016.7533003
[2]   Enhancing Detection Model for Multiple Hypothesis Tracking [J].
Chen, Jiahui ;
Sheng, Hao ;
Zhang, Yang ;
Xiong, Zhang .
2017 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2017, :2143-2152
[3]  
Chen Y., 2019, P CVPR WORKSH
[4]   Deep learning in video multi-object tracking: A survey [J].
Ciaparrone, Gioele ;
Luque Sanchez, Francisco ;
Tabik, Siham ;
Troiano, Luigi ;
Tagliaferri, Roberto ;
Herrera, Francisco .
NEUROCOMPUTING, 2020, 381 :61-88
[5]   ATOM: Accurate Tracking by Overlap Maximization [J].
Danelljan, Martin ;
Bhat, Goutam ;
Khan, Fahad Shahbaz ;
Felsberg, Michael .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :4655-4664
[6]   Siamese Cascaded Region Proposal Networks for Real-Time Visual Tracking [J].
Fan, Heng ;
Ling, Haibin .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :7944-7953
[7]   Graph Convolutional Tracking [J].
Gao, Junyu ;
Zhang, Tianzhu ;
Xu, Changsheng .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :4644-4654
[8]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[9]   Joint detection and online multi-object tracking [J].
Kieritz, Hilke ;
Huebner, Wolfgang ;
Arens, Michael .
PROCEEDINGS 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2018, :1540-1548
[10]   Multi-object Tracking with Neural Gating Using Bilinear LSTM [J].
Kim, Chanho ;
Li, Fuxin ;
Rehg, James M. .
COMPUTER VISION - ECCV 2018, PT VIII, 2018, 11212 :208-224