Low-Latency Video Semantic Segmentation

被引:135
作者
Li, Yule [1 ,4 ]
Shi, Jianping [2 ]
Lin, Dahua [3 ]
机构
[1] Chinese Acad Sci, Inst Comp Technol, Key Lab Intelligent Informat Proc, Beijing 100190, Peoples R China
[2] SenseTime Res, Hong Kong, Peoples R China
[3] Chinese Univ Hong Kong, Dept Informat Engn, Hong Kong, Peoples R China
[4] CUHK Multimedia Lab, Hong Kong, Peoples R China
来源
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR) | 2018年
关键词
D O I
10.1109/CVPR.2018.00628
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent years have seen remarkable progress in semantic segmentation. Yet, it remains a challenging task to apply segmentation techniques to video-based applications. Specifically, the high throughput of video streams, the sheer cost of running fully convolutional networks, together with the low-latency requirements in many real-world applications, e.g. autonomous driving, present a significant challenge to the design of the video segmentation framework. To tackle this combined challenge, we develop a framework for video semantic segmentation, which incorporates two novel components: (1) a feature propagation module that adaptively fuses features over time via spatially variant convolution, thus reducing the cost of per-frame computation; and (2) an adaptive scheduler that dynamically allocate computation based on accuracy prediction. Both components work together to ensure low latency while maintaining high segmentation quality. On both Cityscapes and CamVid, the proposed framework obtained competitive performance compared to the state of the art, while substantially reducing the latency, from 360 ms to 119 ms.
引用
收藏
页码:5997 / 6005
页数:9
相关论文
共 31 条
[1]  
[Anonymous], 2017, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition
[2]  
[Anonymous], 2015, ARXIV150406852
[3]  
[Anonymous], 2015, ABS151100561 CORR
[4]  
[Anonymous], 2016, ARXIV160600915
[5]  
[Anonymous], 2016, ARXIV161105198
[6]  
[Anonymous], ARXIV160807706
[7]  
[Anonymous], 2016, ABS160602147 CORR
[8]  
[Anonymous], 2014, ABS14127062 CORR
[9]  
[Anonymous], ARXIV161208871
[10]  
[Anonymous], 2016, ARXIV161201105CS