FlexPatch: Fast and Accurate Object Detection for On-device High-Resolution Live Video Analytics

被引:25
作者
Yang, Kichang [1 ]
Yi, Juheon [1 ]
Lee, Kyungjin [1 ]
Lee, Youngki [1 ]
机构
[1] Seoul Natl Univ, Seoul, South Korea
来源
IEEE CONFERENCE ON COMPUTER COMMUNICATIONS (IEEE INFOCOM 2022) | 2022年
关键词
live video analytics; On-device AI; high-resolution video; object detection; object tracking;
D O I
10.1109/INFOCOM48880.2022.9796984
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
We present FlexPatch, a novel mobile system to enable accurate and real-time object detection over high-resolution video streams. A widely-used approach for real-time video analysis is detection-based tracking (DBT), i.e., running the heavy-but-accurate detector every few frames and applying a lightweight tracker for in-between frames. However, the approach is limited for real-time processing of high-resolution videos in that i) a lightweight tracker fails to handle occlusion, object appearance changes, and occurrences of new objects, and ii) the detection results do not effectively offset tracking errors due to the high detection latency. We propose tracking-aware patching technique to address such limitations of the DBT frameworks. It effectively identifies a set of subareas where the tracker likely fails and tightly packs them into a small-sized rectangular area where the detection can be efficiently performed at low latency. This prevents the accumulation of tracking errors and offsets the tracking errors with frequent fresh detection results. Our extensive evaluation shows that FlexPatch not only enables real-time and power-efficient analysis of high-resolution frames on mobile devices but also improves the overall accuracy by 146% compared to baseline DBT frameworks.
引用
收藏
页码:1898 / 1907
页数:10
相关论文
共 39 条
[1]   Frugal Following: Power Thrifty Object Detection and Tracking for Mobile Augmented Reality [J].
Apicharttrisorn, Kittipat ;
Ran, Xukan ;
Chen, Jiasi ;
Krishnamurthy, Srikanth, V ;
Roy-Chowdhury, Amit K. .
PROCEEDINGS OF THE 17TH CONFERENCE ON EMBEDDED NETWORKED SENSOR SYSTEMS (SENSYS '19), 2019, :96-109
[2]   Okutama-Action: An Aerial View Video Dataset for Concurrent Human Action Detection [J].
Barekatain, Mohammadamin ;
Marti, Miquel ;
Shih, Hsueh-Fu ;
Murray, Samuel ;
Nakayama, Kotaro ;
Matsuo, Yutaka ;
Prendinger, Helmut .
2017 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2017, :2153-2160
[3]  
Bochkovskiy A, 2020, Arxiv, DOI arXiv:2004.10934
[4]   Cascade R-CNN: Delving into High Quality Object Detection [J].
Cai, Zhaowei ;
Vasconcelos, Nuno .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :6154-6162
[6]   Patchwork: A Patch-wise Attention Network for Efficient Object Detection and Segmentation in Video Streams [J].
Chai, Yuning .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :3414-3423
[7]   MARVEL: Enabling Mobile Augmented Reality with Low Energy and Low Latency [J].
Chen, Kaifei ;
Li, Tong ;
Kim, Hyung-Sin ;
Culler, David E. ;
Katz, Randy H. .
SENSYS'18: PROCEEDINGS OF THE 16TH CONFERENCE ON EMBEDDED NETWORKED SENSOR SYSTEMS, 2018, :292-304
[8]   Glimpse: Continuous, Real-Time Object Recognition on Mobile Devices [J].
Chen, Tiffany Yu-Han ;
Ravindranath, Lenin ;
Deng, Shuo ;
Bahl, Paramvir ;
Balakrishnan, Hari .
SENSYS'15: PROCEEDINGS OF THE 13TH ACM CONFERENCE ON EMBEDDED NETWORKED SENSOR SYSTEMS, 2015, :155-168
[9]  
Chin TW, 2019, Arxiv, DOI arXiv:1902.02910
[10]  
Dai JF, 2016, ADV NEUR IN, V29