User-Generated Video Quality Assessment: A Subjective and Objective Study

被引:21
作者
Li, Yang [1 ]
Meng, Shengbin [4 ]
Zhang, Xinfeng [2 ]
Wang, Meng [3 ]
Wang, Shiqi [3 ]
Wang, Yue [4 ]
Ma, Siwei [1 ]
机构
[1] Peking Univ, Inst Digital Media, Beijing 100871, Peoples R China
[2] Univ Chinese Acad Sci, Sch Comp Sci & Technol, Beijing 101408, Peoples R China
[3] City Univ Hong Kong, Dept Comp Sci, Kwoloon, Hong Kong, Peoples R China
[4] Bytedance Inc, VideoArch Dept, Beijing 100871, Peoples R China
基金
中国国家自然科学基金;
关键词
Deep neural network; user-generated content; video quality assessment; IMAGE; MODEL; INFORMATION; SIMILARITY; EFFICIENT; DATABASE;
D O I
10.1109/TMM.2021.3122347
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Recently, we have observed an exponential increase of user-generated content (UGC) videos. The distinguished characteristic of UGC videos originates from the video production and delivery chain, as they are usually acquired and processed by non-professional users before uploading to the hosting platforms for sharing. As such, these videos usually undergo multiple distortion stages that may affect visual quality before ultimately being viewed. Inspired by the increasing consensus that the optimization of the video coding and processing shall be fully driven by the perceptual quality, in this paper, we propose to study the quality of the UGC videos from both objective and subjective perspectives. We first construct a UGC video quality assessment (VQA) database, aiming to provide useful guidance for the UGC video coding and processing in the hosting platform. The database contains source UGC videos uploaded to the platform and their transcoded versions that are ultimately enjoyed by end-users, along with their subjective scores. Furthermore, we develop an objective quality assessment algorithm that automatically evaluates the quality of the transcoded videos based on the corrupted reference, which is in accordance with the application scenarios of UGC video sharing in the hosting platforms. The information from the corrupted reference is well leveraged and the quality is predicted based on the inferred quality maps with deep neural networks (DNN). Experimental results show that the proposed method yields superior performance. Both subjective and objective evaluations of the UGC videos also shed lights on the design of perceptual UGC video coding.
引用
收藏
页码:154 / 166
页数:13
相关论文
共 67 条
[1]  
[Anonymous], 2009, SUBJ VID QUAL ASS ME
[2]  
[Anonymous], 2012, REC ITU R BT 500 13
[3]   Spatiotemporal Feature Integration and Model Fusion for Full Reference Video Quality Assessment [J].
Bampis, Christos G. ;
Li, Zhi ;
Bovik, Alan C. .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2019, 29 (08) :2256-2270
[4]   SpEED-QA: Spatial Efficient Entropic Differencing for Image and Video Quality [J].
Bampis, Christos G. ;
Gupta, Praful ;
Soundararajan, Rajiv ;
Bovik, Alan C. .
IEEE SIGNAL PROCESSING LETTERS, 2017, 24 (09) :1333-1337
[5]   Deep Neural Networks for No-Reference and Full-Reference Image Quality Assessment [J].
Bosse, Sebastian ;
Maniry, Dominique ;
Mueller, Klaus-Robert ;
Wiegand, Thomas ;
Samek, Wojciech .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2018, 27 (01) :206-219
[6]  
Ffmpeg, About us
[7]   Using multiple spatio-temporal features to estimate video quality [J].
Freitas, Pedro Garcia ;
Akamine, Welington Y. L. ;
Farias, Mylene C. Q. .
SIGNAL PROCESSING-IMAGE COMMUNICATION, 2018, 64 :1-10
[8]   In-Capture Mobile Video Distortions: A Study of Subjective Behavior and Objective Algorithms [J].
Ghadiyaram, Deepti ;
Pan, Janice ;
Bovik, Alan C. ;
Moorthy, Anush Krishna ;
Panda, Prasanjit ;
Yang, Kai-Chieh .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2018, 28 (09) :2061-2077
[9]  
Ghadiyaram D, 2017, IEEE IMAGE PROC, P3445, DOI 10.1109/ICIP.2017.8296922
[10]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778