Multi-step Ahead Visual Trajectory Prediction for Object Tracking using Echo State Networks

被引:0
作者
Manibardo, Eric L. [1 ]
Lana, Ibai [1 ]
Del Ser, Javier [1 ,2 ]
机构
[1] TECNALIA, Basque Res & Technol Alliance, Derio 48160, Bizkaia, Spain
[2] Univ Basque Country UPV EHU, Bilbao 48013, Spain
来源
2023 IEEE 26TH INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS, ITSC | 2023年
关键词
D O I
10.1109/ITSC57777.2023.10422485
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
One of the main applications of multi-object tracking in the context of autonomous driving is improving road safety. An accurate environment understanding where pedestrians and vehicles are correctly identified reduce the risk of an accident while driving. However, occlusions produce identification switches and detection errors, which may involve losing track of an object in the images captured by the vehicle. In this context, tracking by detection is the leading solution. Trackers following this architecture employ a Kalman filter for predicting an object location, encoded as a bounding box within the image boundaries. Having access to a posterior state prediction provides useful information for dealing with occlusions. Unfortunately, the Kalman filter is not designed for producing multi-step ahead predictions. In this work we propose the use of Echo State Networks, (ESN) as a modeling alternative to the Kalman filter. Their recursive nature makes ESNs suited for modeling movement patterns of the bounding boxes detected in the image. Performance results are computed by isolating the motion modules from the tracker itself: a perfect object detector is assumed to enable a detailed analysis of the prediction capabilities of each model over specific object tracks and time slots. Experimental results verify the potential of ESNs for accurate multi-step ahead visual motion prediction. The virtual trajectories delineated by the predicted bounding boxes provide valuable information for anticipating occlusions.
引用
收藏
页码:4782 / 4789
页数:8
相关论文
共 30 条
[1]  
Aharon Nir, 2022, ARXIV220614651
[2]   Tracking without bells and whistles [J].
Bergmann, Philipp ;
Meinhardt, Tim ;
Leal-Taixe, Laura .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :941-951
[3]  
Bewley A, 2016, IEEE IMAGE PROC, P3464, DOI 10.1109/ICIP.2016.7533003
[4]  
Cao Jinkun, 2022, arXiv
[5]  
Dendorfer Patrick., 2020, CoRR abs/2003.09003
[6]   StrongSORT: Make DeepSORT Great Again [J].
Du, Yunhao ;
Zhao, Zhicheng ;
Song, Yang ;
Zhao, Yanyun ;
Su, Fei ;
Gong, Tao ;
Meng, Hongying .
IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 :8725-8737
[7]   GIAOTracker: A comprehensive framework for MCMOT with global information and optimizing strategies in VisDrone 2021 [J].
Du, Yunhao ;
Wan, Junfeng ;
Zhao, Yanyun ;
Zhang, Binyu ;
Tong, Zhihang ;
Dong, Junhao .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW 2021), 2021, :2809-2819
[8]  
Ge Z., 2021, ABS210708430 CORR, DOI [10.48550/arXiv.2107.08430, 10.48550/ARXIV.2107.08430]
[9]  
Geiger A, 2012, PROC CVPR IEEE, P3354, DOI 10.1109/CVPR.2012.6248074
[10]  
Henschel R., 2019, IEEE CVF C COMP VIS, P0