LiveDeep: Online Viewport Prediction for Live Virtual Reality Streaming Using Lifelong Deep Learning

被引:0
作者
Feng, Xianglong [1 ]
Liu, Yao [2 ]
Wei, Sheng [1 ]
机构
[1] Rutgers State Univ, Newark, NJ 07102 USA
[2] SUNY Binghamton, Binghamton, NY 13902 USA
来源
2020 IEEE CONFERENCE ON VIRTUAL REALITY AND 3D USER INTERFACES (VR 2020) | 2020年
基金
美国国家科学基金会;
关键词
Human-centered computing; Human computer interaction (HCI); Interaction paradigms; Virtual reality;
D O I
10.1109/VR46266.2020.00005
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Live virtual reality (VR) streaming has become a popular and trending video application in the consumer market providing users with 360-degree, immersive viewing experiences. To provide premium quality of experience, VR streaming faces unique challenges due to the significantly increased bandwidth consumption. To address the bandwidth challenge, VR video viewport prediction has been proposed as a viable solution, which predicts and streams only the user's viewport of interest with high quality to the VR device. However, most of the existing viewport prediction approaches target only the video-on-demand (VOD) use cases, requiring offline processing of the historical video and/or user data that are not available in the live streaming scenario. In this work, we develop a novel viewport prediction approach for live VR streaming, which only requires video content and user data in the current viewing session. To address the challenges of insufficient training data and real-time processing, we propose a live VR-specific deep learning mechanism, namely LiveDeep, to create the online viewport prediction model and conduct real-time inference. LiveDeep employs a hybrid approach to address the unique challenges in live VR streaming, involving (1) an alternate online data collection, labeling, training, and inference schedule with controlled feedback loop to accommodate for the sparse training data; and (2) a mixture of hybrid neural network models to accommodate for the inaccuracy caused by a single model. We evaluate LiveDeep using 48 users and 14 VR videos of various types obtained from a public VR user head movement dataset. The results indicate around 90% prediction accuracy, around 40% bandwidth savings, and premium processing time, which meets the bandwidth and real-time requirements of live VR streaming.
引用
收藏
页码:800 / 808
页数:9
相关论文
共 42 条
[1]  
[Anonymous], ACM WORKSH NETW OP S
[2]  
[Anonymous], CLOUD AR VR WHIT
[3]  
[Anonymous], VR AR WILL PUSH LIMI
[4]  
[Anonymous], HTC VIVE REV
[5]  
[Anonymous], AR VR HEADS RET GROW
[6]  
[Anonymous], FACEBOOK NOW LETS YO
[7]  
[Anonymous], 2019, ABC NEWS
[8]  
[Anonymous], 2019, ARXIV190201439
[9]  
[Anonymous], WATCH AN YOUTUBE VID
[10]  
[Anonymous], REPOSITORY PROJECT L