PGCNet: patch graph convolutional network for point cloud segmentation of indoor scenes

被引:22
作者
Sun, Yuliang [1 ]
Miao, Yongwei [2 ]
Chen, Jiazhou [1 ]
Pajarola, Renato [3 ]
机构
[1] Zhejiang Univ Technol, Coll Comp Sci & Technol, Hangzhou, Peoples R China
[2] Zhejiang Sci Tech Univ, Coll Informat Sci & Technol, Hangzhou, Peoples R China
[3] Univ Zurich, Dept Informat, CH-8050 Zurich, Switzerland
基金
中国国家自然科学基金;
关键词
Point cloud; Scene segmentation; Surface patch; Graph convolutional network; Edge convolution; Encoder-decoder;
D O I
10.1007/s00371-020-01892-8
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Semantic segmentation of 3D point clouds is a crucial task in scene understanding and is also fundamental to indoor scene applications such as indoor navigation, mobile robotics, augmented reality. Recently, deep learning frameworks have been successfully adopted to point clouds but are limited by the size of data. While most existing works focus on individual sampling points, we use surface patches as a more efficient representation and propose a novel indoor scene segmentation framework called patch graph convolution network (PGCNet). This framework treats patches as input graph nodes and subsequently aggregates neighboring node features by dynamic graph U-Net (DGU) module, which consists of dynamic edge convolution operation inside U-shaped encoder-decoder architecture. The DGU module dynamically update graph structures at each level to encode hierarchical edge features. Incorporating PGCNet, we can segment the input scene into two types, i.e., room layout and indoor objects, which is afterward utilized to carry out final rich semantic labeling of various indoor scenes. With considerable speedup training, the proposed framework achieves effective performance equivalent to state-of-the-art for segmenting standard indoor scene dataset.
引用
收藏
页码:2407 / 2418
页数:12
相关论文
共 43 条
[1]  
[Anonymous], 2016, PROC CVPR IEEE, DOI DOI 10.1109/CVPR.2016.170
[2]   Shape Completion using 3D-Encoder-Predictor CNNs and Shape Synthesis [J].
Dai, Angela ;
Qi, Charles Ruizhongtai ;
Niessner, Matthias .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :6545-6554
[3]   ScanNet: Richly-annotated 3D Reconstructions of Indoor Scenes [J].
Dai, Angela ;
Chang, Angel X. ;
Savva, Manolis ;
Halber, Maciej ;
Funkhouser, Thomas ;
Niessner, Matthias .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :2432-2443
[4]  
Defferrard M., 2016, Advances in Neural Information Processing Systems, DOI DOI 10.5555/3157382.3157527
[5]  
Gao HY, 2019, PR MACH LEARN RES, V97
[6]   Correspondence estimation for non-rigid point clouds with automatic part discovery [J].
Guo, Hao ;
Zhu, Dehai ;
Mordohai, Philippos .
VISUAL COMPUTER, 2016, 32 (12) :1511-1524
[7]   Point-wise saliency detection on 3D point clouds via covariance descriptors [J].
Guo, Yu ;
Wang, Fei ;
Xin, Jingmin .
VISUAL COMPUTER, 2018, 34 (10) :1325-1338
[8]   3D Object Recognition in Cluttered Scenes with Local Surface Features: A Survey [J].
Guo, Yulan ;
Bennamoun, Mohammed ;
Sohel, Ferdous ;
Lu, Min ;
Wan, Jianwei .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2014, 36 (11) :2270-2287
[9]   3D-SIS: 3D Semantic Instance Segmentation of RGB-D Scans [J].
Hou, Ji ;
Dai, Angela ;
Niessner, Matthias .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :4416-4425
[10]   Semantic Labeling and Instance Segmentation of 3D Point Clouds Using Patch Context Analysis and Multiscale Processing [J].
Hu, Shi-Min ;
Cai, Jun-Xiong ;
Lai, Yu-Kun .
IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2020, 26 (07) :2485-2498