Subjective and Objective Quality Assessment of Rendered Human Avatar Videos in Virtual Reality

被引:0
作者
Chen, Yu-Chih [1 ]
Saha, Avinab [1 ]
Chapiro, Alexandre [2 ]
Hane, Christian [2 ]
Bazin, Jean-Charles [2 ]
Qiu, Bo [2 ]
Zanetti, Stefano [2 ]
Katsavounidis, Ioannis [2 ]
Bovik, Alan C. [1 ]
机构
[1] Univ Texas Austin, Dept Elect & Comp Engn, Lab Image & Video Engn LIVE, Austin, TX 94025 USA
[2] Meta Platforms Inc, Menlo Pk, CA 94025 USA
基金
美国国家科学基金会;
关键词
Avatars; Videos; Three-dimensional displays; Quality assessment; Monitoring; Databases; Predictive models; Visualization; Solid modeling; Image coding; Virtual reality; video quality assessment; 3D mesh; human avatar video; six degrees of freedom; VISUAL QUALITY; POINT CLOUDS; MESH; INFORMATION;
D O I
10.1109/TIP.2024.3468881
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We study the visual quality judgments of human subjects on digital human avatars (sometimes referred to as "holograms" in the parlance of virtual reality [VR] and augmented reality [AR] systems) that have been subjected to distortions. We also study the ability of video quality models to predict human judgments. As streaming human avatar videos in VR or AR become increasingly common, the need for more advanced human avatar video compression protocols will be required to address the tradeoffs between faithfully transmitting high-quality visual representations while adjusting to changeable bandwidth scenarios. During transmission over the internet, the perceived quality of compressed human avatar videos can be severely impaired by visual artifacts. To optimize trade-offs between perceptual quality and data volume in practical workflows, video quality assessment (VQA) models are essential tools. However, there are very few VQA algorithms developed specifically to analyze human body avatar videos, due, at least in part, to the dearth of appropriate and comprehensive datasets of adequate size. Towards filling this gap, we introduce the LIVE-Meta Rendered Human Avatar VQA Database, which contains 720 human avatar videos processed using 20 different combinations of encoding parameters, labeled by corresponding human perceptual quality judgments that were collected in six degrees of freedom VR headsets. To demonstrate the usefulness of this new and unique video resource, we use it to study and compare the performances of a variety of state-of-the-art Full Reference and No Reference video quality prediction models, including a new model called HoloQA. As a service to the research community, we publicly releases the metadata of the new database at https://live.ece.utexas.edu/research/LIVE-Meta-rendered-human-avatar/index.html.
引用
收藏
页码:5740 / 5754
页数:15
相关论文
共 80 条
[21]  
Gutierrez J., 2020, Electron. Imag., V32
[22]   Point Cloud Rendering After Coding: Impacts on Subjective and Objective Quality [J].
Javaheri, Alireza ;
Brites, Catarina ;
Pereira, Fernando ;
Ascenso, Joao .
IEEE TRANSACTIONS ON MULTIMEDIA, 2021, 23 :4049-4064
[23]  
Javaheri A, 2017, IEEE INT WORKSH MULT
[24]  
Jocher G., 2023, YOLO by Ultralytics
[26]   On the Efficiency of Image Metrics for Evaluating the Visual Quality of 3D Models [J].
Lavoue, Guillaume ;
Larabi, Mohamed Chaker ;
Vasa, Libor .
IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2016, 22 (08) :1987-1999
[27]   A Multiscale Metric for 3D Mesh Visual Quality Assessment [J].
Lavoue, Guillaume .
COMPUTER GRAPHICS FORUM, 2011, 30 (05) :1427-1437
[28]   Quality Assessment of In-the-Wild Videos [J].
Li, Dingquan ;
Jiang, Tingting ;
Jiang, Ming .
PROCEEDINGS OF THE 27TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA (MM'19), 2019, :2351-2359
[29]   Image Quality Assessment by Separately Evaluating Detail Losses and Additive Impairments [J].
Li, Songnan ;
Zhang, Fan ;
Ma, Lin ;
Ngan, King Ngi .
IEEE TRANSACTIONS ON MULTIMEDIA, 2011, 13 (05) :935-949
[30]  
Li Z., 2019, P 17 INT C VIRT REAL, P1