Subjective and Objective Quality Assessment of Rendered Human Avatar Videos in Virtual Reality

被引:0
作者
Chen, Yu-Chih [1 ]
Saha, Avinab [1 ]
Chapiro, Alexandre [2 ]
Hane, Christian [2 ]
Bazin, Jean-Charles [2 ]
Qiu, Bo [2 ]
Zanetti, Stefano [2 ]
Katsavounidis, Ioannis [2 ]
Bovik, Alan C. [1 ]
机构
[1] Univ Texas Austin, Dept Elect & Comp Engn, Lab Image & Video Engn LIVE, Austin, TX 94025 USA
[2] Meta Platforms Inc, Menlo Pk, CA 94025 USA
基金
美国国家科学基金会;
关键词
Avatars; Videos; Three-dimensional displays; Quality assessment; Monitoring; Databases; Predictive models; Visualization; Solid modeling; Image coding; Virtual reality; video quality assessment; 3D mesh; human avatar video; six degrees of freedom; VISUAL QUALITY; POINT CLOUDS; MESH; INFORMATION;
D O I
10.1109/TIP.2024.3468881
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We study the visual quality judgments of human subjects on digital human avatars (sometimes referred to as "holograms" in the parlance of virtual reality [VR] and augmented reality [AR] systems) that have been subjected to distortions. We also study the ability of video quality models to predict human judgments. As streaming human avatar videos in VR or AR become increasingly common, the need for more advanced human avatar video compression protocols will be required to address the tradeoffs between faithfully transmitting high-quality visual representations while adjusting to changeable bandwidth scenarios. During transmission over the internet, the perceived quality of compressed human avatar videos can be severely impaired by visual artifacts. To optimize trade-offs between perceptual quality and data volume in practical workflows, video quality assessment (VQA) models are essential tools. However, there are very few VQA algorithms developed specifically to analyze human body avatar videos, due, at least in part, to the dearth of appropriate and comprehensive datasets of adequate size. Towards filling this gap, we introduce the LIVE-Meta Rendered Human Avatar VQA Database, which contains 720 human avatar videos processed using 20 different combinations of encoding parameters, labeled by corresponding human perceptual quality judgments that were collected in six degrees of freedom VR headsets. To demonstrate the usefulness of this new and unique video resource, we use it to study and compare the performances of a variety of state-of-the-art Full Reference and No Reference video quality prediction models, including a new model called HoloQA. As a service to the research community, we publicly releases the metadata of the new database at https://live.ece.utexas.edu/research/LIVE-Meta-rendered-human-avatar/index.html.
引用
收藏
页码:5740 / 5754
页数:15
相关论文
共 80 条
[31]  
Li Z, 2021, Arxiv, DOI [arXiv:2004.02067, DOI 10.2352/ISSN.2470-1173.2020.11.HVEI-131]
[32]   Recover Subjective Quality Scores from Noisy Measurements [J].
Li, Zhi ;
Bampis, Christos G. .
2017 DATA COMPRESSION CONFERENCE (DCC), 2017, :52-61
[33]  
Li Zhi, 2016, Netflix Technology Blog
[34]   Point Cloud Quality Assessment: Dataset Construction and Learning-based No-reference Metric [J].
Liu, Yipeng ;
Yang, Qi ;
Xu, Yiling ;
Yang, Le .
ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2023, 19 (02)
[35]   CONVIQT: Contrastive Video Quality Estimator [J].
Madhusudana, Pavan C. ;
Birkbeck, Neil ;
Wang, Yilin ;
Adsumilli, Balu ;
Bovik, Alan C. .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2023, 32 :5138-5152
[36]   Image Quality Assessment Using Contrastive Learning [J].
Madhusudana, Pavan C. ;
Birkbeck, Neil ;
Wang, Yilin ;
Adsumilli, Balu ;
Bovik, Alan C. .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2022, 31 :4149-4161
[37]   ST-GREED: Space-Time Generalized Entropic Differences for Frame Rate Dependent Video Quality Prediction [J].
Madhusudana, Pavan C. ;
Birkbeck, Neil ;
Wang, Yilin ;
Adsumilli, Balu ;
Bovik, Alan C. .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 :7446-7457
[38]   FovVideoVDP: A visible difference predictor for wide field-of-view video [J].
Mantiuk, Rafal K. ;
Denes, Gyorgy ;
Chapiro, Alexandre ;
Kaplanyan, Anton ;
Rufo, Gizem ;
Bachy, Romain ;
Lian, Trisha ;
Patney, Anjul .
ACM TRANSACTIONS ON GRAPHICS, 2021, 40 (04)
[39]   A "beyond being there" for VR meetings: envisioning the future of remote work [J].
McVeigh-Schultz, Joshua ;
Isbister, Katherine .
HUMAN-COMPUTER INTERACTION, 2022, 37 (05) :433-453
[40]   Design, Implementation, and Evaluation of a Point Cloud Codec for Tele-Immersive Video [J].
Mekuria, Rufael ;
Blom, Kees ;
Cesar, Pablo .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2017, 27 (04) :828-842