DCFNet:Dual-Channel Feature Fusion of Real Scene for Point Cloud Semantic Segmentation

被引:0
作者
Sun, Liujie [1 ]
Zhu, Yaoda [1 ]
Wang, Wenju [1 ]
机构
[1] College of Communication and Art Design, Shanghai University of Science and Technology, Shanghai
关键词
attention mechanism; deep learning; dual-channel feature fusion; point cloud semantic segmentation;
D O I
10.3778/j.issn.1002-8331.2305-0290
中图分类号
学科分类号
摘要
The point cloud of the real scene not only has the spatial geometric information of the point cloud, but also has the color information of the 3D object. The existing network cannot effectively use the local features and spatial geometric feature information of the real scene. Therefore, a dual-channel feature fusion of real scene for point cloud semantic segmentation DCFNet can be used for indoor and outdoor scene semantic segmentation in different scenarios. More specifically, in order to solve the problem that the color information of the point cloud in the real scene cannot be fully extracted, the method uses two input channels, and the channel adopts the same feature extraction network structure. The input of the upper channel is the complete RGB color and point cloud coordinate information, and the channel mainly focuses on the scene features of complex objects, while the lower channel only inputs the point cloud coordinate information. This channel mainly focuses on the spatial geometric characteristics of the point cloud. In each channel, in order to better extract local and global information and improve network performance, the inter-layer fusion module and the Transformer channel feature expansion module are introduced. At the same time, the existing 3D point cloud semantic segmentation methods lack of attention to the relationship between local features and global features, which leads to poor segmentation results for complex scenes. In this paper, the features extracted from the upper and lower channels are fused by the DCFFS (dual-channel feature fusion segmentation) module, and the semantic segmentation of the real scene is performed. The experimental results show that the mean intersection over union (MIOU) of the proposed DCFNet segmentation method on the S3DIS Area5 indoor scene dataset and the STPLS3D outdoor scene dataset reaches 71.18% and 48.87% respectively. The mean average precision (MACC) and overall accuracy (OACC) reach 77.01% and 86.91% respectively, achieving high-precision point cloud semantic segmentation in real scenes. © 2024 Journal of Computer Engineering and Applications Beijing Co., Ltd.; Science Press. All rights reserved.
引用
收藏
页码:160 / 169
页数:9
相关论文
共 30 条
[1]  
GUO M H, CAI J X, LIU Z N, Et al., PCT: point cloud transformer, Computational Visual Media, 7, 2, pp. 187-199, (2021)
[2]  
ZHAO H S, JIANG L, JIA J Y, Et al., Point transformer, 18th IEEE/CVF International Conference on Computer Vision (ICCV), pp. 16239-16248, (2021)
[3]  
ZHANG R, ZENG Z, GUO Z, Et al., DSPoint: dual-scale point cloud recognition with high-frequency fusion, (2021)
[4]  
RAN H, LIU J, WANG C., Surface representation for point clouds, Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 18942-18952, (2022)
[5]  
HU Q Y, YANG B, XIE L H, Et al., RandLA-Net: efficient semantic segmentation of large-scale point clouds, Proceedings 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 11105-11114, (2020)
[6]  
LIU K, GAO Z, LIN F, Et al., FG-Net: a fast and accurate framework for large-scale lidar point cloud understanding, IEEE Transactions on Cybernetics, 53, 1, pp. 553-564, (2022)
[7]  
RAN H X, ZHUO W, LIU J, Et al., Learning inner-group relations on point clouds, 18th IEEE/CVF International Conference on Computer Vision (ICCV), pp. 15457-15467, (2021)
[8]  
THOMAS H, QI C R, DESCHAUD J E, Et al., KPConv: flexible and deformable convolution for point clouds, IEEE/ CVF International Conference on Computer Vision (ICCV), pp. 6420-6429, (2019)
[9]  
GUO Y, WANG H, HU Q, Et al., Deep learning for 3d point clouds: a survey, IEEE Transactions on Pattern Analysis and Machine Intelligence, 43, 12, pp. 4338-4364, (2020)
[10]  
LI Y, BU R, SUN M, Et al., PointCNN: convolution on Χ-transformed points, Proceedings of the 32nd International Conference on Neural Information Processing Systems, pp. 828-838, (2018)