TrackNet: A Deep Learning Network for Tracking High-speed and Tiny Objects in Sports Applications

被引:33
作者
Huang, Yu-Chuan [1 ]
Liao, I-No [1 ]
Chen, Ching-Hsuan [1 ]
Ik, Ts I-Ui [1 ]
Peng, Wen-Chih [1 ]
机构
[1] Natl Chiao Tung Univ, Dept Comp Sci, 1001 Univ Rd, Hsinchu 30010, Taiwan
来源
2019 16TH IEEE INTERNATIONAL CONFERENCE ON ADVANCED VIDEO AND SIGNAL BASED SURVEILLANCE (AVSS) | 2019年
关键词
D O I
10.1109/avss.2019.8909871
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Ball trajectory data are one of the most fundamental and useful information in the evaluation of players' performance and analysis of game strategies. It is still challenging to recognize and position a high-speed and tiny ball accurately from an ordinary video. In this paper, we develop a deep learning network, called TrackNet, to track the tennis ball from broadcast videos in which the ball images are small, blurry, and sometimes with afterimage tracks or even invisible. The proposed heatmap-based deep learning network is trained to not only recognize the ball image from a single frame but also learn flying patterns from consecutive frames. The network is evaluated on the video of the men's singles final at the 2017 Summer Universiade, which is available on YouTube. The precision, recall, and F1-measure reach 99:7%, 97:3%, and 98:5%, respectively. To prevent overfitting, 9 additional videos are partially labeled together with a subset from the previous dataset to implement 10-fold cross-validation, and the precision, recall, and F1-measure are 95:3%, 75:7%, and 84:3%, respectively.
引用
收藏
页数:8
相关论文
共 19 条
[1]   Object Detection and Tracking based on Trajectory in Broadcast Tennis Video [J].
Archana, M. ;
Geetha, M. Kalaisevi .
SECOND INTERNATIONAL SYMPOSIUM ON COMPUTER VISION AND THE INTERNET (VISIONNET'15), 2015, 58 :225-232
[2]   Recurrent Human Pose Estimation [J].
Belagiannis, Vasileios ;
Zisserman, Andrew .
2017 12TH IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION (FG 2017), 2017, :468-475
[3]   Ball tracking and 3D trajectory approximation with applications to tactics analysis from single-camera volleyball sequences [J].
Chen, Hua-Tsung ;
Tsai, Wen-Jiin ;
Lee, Suh-Yin ;
Yu, Jen-Yu .
MULTIMEDIA TOOLS AND APPLICATIONS, 2012, 60 (03) :641-667
[4]   A Deep Learning Approach to Human Activity Recognition Based on Single Accelerometer [J].
Chen, Yuqing ;
Xue, Yang .
2015 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC 2015): BIG DATA ANALYTICS FOR HUMAN-CENTRIC SYSTEMS, 2015, :1488-1492
[5]  
Fu T.-S., 2011, PROC 2011 IEEE VIS C
[6]   Human Activity Recognition using Wearable Sensors by Deep Convolutional Neural Networks [J].
Jiang, Wenchao ;
Yin, Zhaozheng .
MM'15: PROCEEDINGS OF THE 2015 ACM MULTIMEDIA CONFERENCE, 2015, :1307-1310
[7]   ImageNet Classification with Deep Convolutional Neural Networks [J].
Krizhevsky, Alex ;
Sutskever, Ilya ;
Hinton, Geoffrey E. .
COMMUNICATIONS OF THE ACM, 2017, 60 (06) :84-90
[8]  
Long J, 2015, PROC CVPR IEEE, P3431, DOI 10.1109/CVPR.2015.7298965
[9]  
Myint H, 2015, 2015 14TH IAPR INTERNATIONAL CONFERENCE ON MACHINE VISION APPLICATIONS (MVA), P170, DOI 10.1109/MVA.2015.7153160
[10]   Learning Deconvolution Network for Semantic Segmentation [J].
Noh, Hyeonwoo ;
Hong, Seunghoon ;
Han, Bohyung .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :1520-1528