MFFAE-Net: semantic segmentation of point clouds using multi-scale feature fusion and attention enhancement networks

被引:0
作者
Liu, Wei [1 ,2 ,3 ]
Lu, Yisheng [1 ,2 ]
Zhang, Tao [1 ,2 ]
机构
[1] China Univ Geosci, Sch Automat, Wuhan 430074, Hubei, Peoples R China
[2] China Univ Geosci, Hubei Key Lab Adv Control & Intelligent Automat, Wuhan 430074, Hubei, Peoples R China
[3] Minist Educ, Engn Res Ctr Intelligent Technol Geoexplorat, Wuhan 430074, Hubei, Peoples R China
关键词
Semantic segmentation; Feature enhancement; Feature fusion; Point cloud; CLASSIFICATION;
D O I
10.1007/s00138-024-01589-1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Point cloud data can reflect more information about the real 3D space, which has gained increasing attention in computer vision field. But the unstructured and unordered nature of point clouds poses many challenges in their study. How to learn the global features of the point cloud in the original point cloud is a problem that has been accompanied by the research. In the research based on the structure of the encoder and decoder, many researchers focus on designing the encoder to better extract features, and do not further explore more globally representative features according to the features of the encoder and decoder. To solve this problem, we propose the MFFAE-Net method, which aims to obtain more globally representative point cloud features by using the feature learning of encoder decoder stage.Our method first enhances the feature information of the input point cloud by merging the information of its neighboring points, which is helpful for the following point cloud feature extraction work. Secondly, the channel attention module is used to further process the extracted features, so as to highlight the role of important channels in the features. Finally, we fuse features of different scales from encoding features and decoding features as well as features of the same scale, so as to obtain more global point cloud features, which will help improve the segmentation results of point clouds. Experimental results show that the method performs well on some objects in S3DIS dataset and Toronto3d dataset.
引用
收藏
页数:15
相关论文
共 38 条
[1]   3D Semantic Parsing of Large-Scale Indoor Spaces [J].
Armeni, Iro ;
Sener, Ozan ;
Zamir, Amir R. ;
Jiang, Helen ;
Brilakis, Ioannis ;
Fischer, Martin ;
Savarese, Silvio .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :1534-1543
[2]   SnapNet: 3D point cloud semantic labeling with 2D deep segmentation networks [J].
Boulch, Alexandre ;
Guerry, Yids ;
Le Saux, Bertrand ;
Audebert, Nicolas .
COMPUTERS & GRAPHICS-UK, 2018, 71 :189-198
[3]  
Chen K., 2022, ARXIV
[4]   3D Point Cloud Classification for Autonomous Driving via Dense-Residual Fusion Network [J].
Chiang, Chung-Hsin ;
Kuo, Chih-Hung ;
Lin, Chien-Chou ;
Chiang, Hsin-Te .
IEEE ACCESS, 2020, 8 :163775-163783
[5]   Shape Completion using 3D-Encoder-Predictor CNNs and Shape Synthesis [J].
Dai, Angela ;
Qi, Charles Ruizhongtai ;
Niessner, Matthias .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :6545-6554
[6]   Semi-automatic Extraction of Rural Roads From High-Resolution Remote Sensing Images Based on a Multifeature Combination [J].
Dai, Jiguang ;
Ma, Rongchen ;
Ai, Haibin .
IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2022, 19
[7]   3-D Mapping With an RGB-D Camera [J].
Endres, Felix ;
Hess, Juergen ;
Sturm, Juergen ;
Cremers, Daniel ;
Burgard, Wolfram .
IEEE TRANSACTIONS ON ROBOTICS, 2014, 30 (01) :177-187
[8]   SCF-Net: Learning Spatial Contextual Features for Large-Scale Point Cloud Segmentation [J].
Fan, Siqi ;
Dong, Qiulei ;
Zhu, Fenghua ;
Lv, Yisheng ;
Ye, Peijun ;
Wang, Fei-Yue .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :14499-14508
[9]   Point-cloud based 3D object detection and classification methods for self-driving applications: A survey and taxonomy [J].
Fernandes, Duarte ;
Silva, Antonio ;
Nevoa, Rafael ;
Simoes, Claudia ;
Gonzalez, Dibet ;
Guevara, Miguel ;
Novais, Paulo ;
Monteiro, Joao ;
Melo-Pinto, Pedro .
INFORMATION FUSION, 2021, 68 :161-191
[10]  
Hu J, 2018, PROC CVPR IEEE, P7132, DOI [10.1109/CVPR.2018.00745, 10.1109/TPAMI.2019.2913372]