Video-Based Point Cloud Compression Artifact Removal

被引:35
作者
Akhtar, Anique [1 ]
Gao, Wen [2 ,3 ]
Li, Li [1 ,3 ]
Li, Zhu [1 ]
Jia, Wei [1 ]
Liu, Shan [2 ,3 ]
机构
[1] Univ Missouri Kansas City, Dept Comp Sci & Elect Engn, Kansas City, MO 64110 USA
[2] Tencent Amer, Palo Alto, CA 94301 USA
[3] Univ Sci & Technol China, CAS, Key Lab Technol Geospatial Informat Proc & Applic, Hefei 230027, Peoples R China
关键词
Three-dimensional displays; Geometry; Quantization (signal); Bit rate; Encoding; Vehicle dynamics; Deep learning; Point cloud; artifact removal; compression; 3D deep learning; quantization; V-PCC; MPEG;
D O I
10.1109/TMM.2021.3090148
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Photo-realistic point cloud capture and transmission are the fundamental enablers for immersive visual communication. The coding process of dynamic point clouds, especially video-based point cloud compression (V-PCC) developed by the MPEG standardization group, is now delivering state-of-the-art performance in compression efficiency. V-PCC is based on the projection of the point cloud patches to 2D planes and encoding the sequence as 2D texture and geometry patch sequences. However, the resulting quantization errors from coding can introduce compression artifacts, which can be very unpleasant for the quality of experience (QoE). In this work, we developed a novel out-of-the-loop point cloud geometry artifact removal solution that can significantly improve reconstruction quality without additional bandwidth cost. Our novel framework consists of a point cloud sampling scheme, an artifact removal network, and an aggregation scheme. The point cloud sampling scheme employs a cube-based neighborhood patch extraction to divide the point cloud into patches. The geometry artifact removal network then processes these patches to obtain artifact-removed patches. The artifact-removed patches are then merged together using an aggregation scheme to obtain the final artifact-removed point cloud. We employ 3D deep convolutional feature learning for geometry artifact removal that jointly recovers both the quantization direction and the quantization noise level by exploiting projection and quantization prior. The simulation results demonstrate that the proposed method is highly effective and can considerably improve the quality of the reconstructed point cloud.
引用
收藏
页码:2866 / 2876
页数:11
相关论文
共 51 条
[1]  
Akhtar A., 2019, P ICC IEEE INT C COM, P1
[2]  
Akhtar A, 2019, IEEE IMAGE PROC, P2369, DOI [10.1109/ICIP.2019.8803373, 10.1109/icip.2019.8803373]
[3]  
[Anonymous], ISO/IEC JTC1/SC29/WG11, call for proposals on 3d video coding technology, Doc
[4]  
[Anonymous], 2017, Standard N16763
[5]  
[Anonymous], 2014, 6DoF vs 3DoF
[6]  
[Anonymous], 2019, MOBILE MAPPING SYSTE
[7]  
[Anonymous], DOCUMENT ISOIEC JTC1
[8]  
Bjontegaard G., 2001, ITU-T SG.16 Q.6 (VCEG), VCEG-M33
[9]   4D Spatio-Temporal ConvNets: Minkowski Convolutional Neural Networks [J].
Choy, Christopher ;
Gwak, JunYoung ;
Savarese, Silvio .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :3070-3079
[10]   Image denoising by sparse 3-D transform-domain collaborative filtering [J].
Dabov, Kostadin ;
Foi, Alessandro ;
Katkovnik, Vladimir ;
Egiazarian, Karen .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2007, 16 (08) :2080-2095