Observation-Centric with Appearance Metric for Computer Vision-Based Vehicle Counting

被引:1
作者
Brillantes, Allysa Kate [1 ,2 ]
Sybingco, Edwin [1 ,2 ,3 ]
Billones, Robert Kerwin [2 ,3 ,4 ]
Bandala, Argel [1 ,2 ]
Fillone, Alexis [2 ,3 ,5 ]
Dadios, Elmer [2 ,3 ,4 ]
机构
[1] De La Salle Univ, Dept Elect & Comp Engn, Manila, Philippines
[2] Ctr Engn & Sustainable Dev Res, Manila, Philippines
[3] Intelligent Syst Innovat ISI Inc, Pasig, Philippines
[4] De La Salle Univ, Dept Mfg Engn & Management, Manila, Philippines
[5] De La Salle Univ, Dept Civil Engn, Manila, Philippines
关键词
mask-OCSORT; object tracking; instance segmentation; traffic surveillance systems; low framerate; data association; TRACKING;
D O I
10.12720/jait.14.6.1261-1272
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Tracking objects in video sequences is a key step for applications involving computer vision like traffic monitoring and security systems. Occlusion is a frequent problem in object tracking which can result in the tracker losing track of the occluded object or misidentifying it with the occluding object. Moreover, the limited memory and computing power of traffic analysis systems presents a scaling problem, especially in object tracking applications. This paper aims to improve object tracking performance by minimizing data association errors in low frame rate tracking applications. Reducing frame rates alleviates memory and computing power limitations, and utilizing a tracker that can handle occlusion can address occlusion-related issues in object tracking. The proposed tracking method, Mask-OCSORT, uses the observation-tracking method with cosine similarity, intersection-over-union, and velocity consistency metrics for the association problem. The paper analyzes the effect of using bounding box and mask predictions of deep learning models in generating tracks. This study uses evaluation metrics like HOTA, MOTA, and IDF1 to assess the proposed tracking method, and employs evaluation metrics such as precision, recall, and F-score to assess the counting based on generated IDs from the tracking method. The study applied the Mask-OCSORT tracking for vehicle counting application and achieved an F-score of 87.18% at 5 frames per second (fps), and 75% at 1 fps.
引用
收藏
页码:1261 / 1272
页数:12
相关论文
共 47 条
[1]  
Bewley A, 2016, IEEE IMAGE PROC, P3464, DOI 10.1109/ICIP.2016.7533003
[2]   YOLACT Real-time Instance Segmentation [J].
Bolya, Daniel ;
Zhou, Chong ;
Xiao, Fanyi ;
Lee, Yong Jae .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :9156-9165
[3]  
Cao JK, 2023, Arxiv, DOI arXiv:2203.14360
[4]   YolTrack: Multitask Learning Based Real-Time Multiobject Tracking and Segmentation for Autonomous Vehicles [J].
Chang, Xuepeng ;
Pan, Huihui ;
Sun, Weichao ;
Gao, Huijun .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (12) :5323-5333
[5]   An Edge Traffic Flow Detection Scheme Based on Deep Learning in an Intelligent Transportation System [J].
Chen, Chen ;
Liu, Bin ;
Wan, Shaohua ;
Qiao, Peng ;
Pei, Qingqi .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2021, 22 (03) :1840-1852
[6]  
Chen L, 2017, IEEE IMAGE PROC, P645, DOI 10.1109/ICIP.2017.8296360
[7]  
Chen M., 2022, arXiv
[8]   Deep-Learning Based Joint Iris and Sclera Recognition with YOLO Network for Identity Identification [J].
Chuang, Chia-Wei ;
Fan, Chih-Peng .
JOURNAL OF ADVANCES IN INFORMATION TECHNOLOGY, 2021, 12 (01) :60-65
[9]   Deep learning in video multi-object tracking: A survey [J].
Ciaparrone, Gioele ;
Luque Sanchez, Francisco ;
Tabik, Siham ;
Troiano, Luigi ;
Tagliaferri, Roberto ;
Herrera, Francisco .
NEUROCOMPUTING, 2020, 381 :61-88
[10]  
Dai P., 2022, arXiv