An improved fused feature residual network for 3D point cloud data

被引:0
作者
Gezawa, Abubakar Sulaiman [1 ]
Liu, Chibiao [1 ]
Jia, Heming [1 ]
Nanehkaran, Y. A. [2 ]
Almutairi, Mubarak S. [3 ]
Chiroma, Haruna [4 ]
机构
[1] Sanming Univ, Coll Informat Engn, Fujian Key Lab Agr IOT Applicat, Sanming, Fujian, Peoples R China
[2] Yancheng Teachers Univ, Sch Informat Engn, Dept Software Engn, Yancheng, Jiangsu, Peoples R China
[3] Univ Hafr Al Batin, Coll Comp Sci & Engn, Hafar al Batin, Saudi Arabia
[4] Univ Hafr Al Batin, Coll Comp Sci & Engn Technol, Appl Coll, Hafar Al Batin, Saudi Arabia
关键词
point clouds; part segmentation; classification; shape features; 3D objects recognition; CLASSIFICATION; SEGMENTATION;
D O I
10.3389/fncom.2023.1204445
中图分类号
Q [生物科学];
学科分类号
07 ; 0710 ; 09 ;
摘要
Point clouds have evolved into one of the most important data formats for 3D representation. It is becoming more popular as a result of the increasing affordability of acquisition equipment and growing usage in a variety of fields. Volumetric grid-based approaches are among the most successful models for processing point clouds because they fully preserve data granularity while additionally making use of point dependency. However, using lower order local estimate functions to close 3D objects, such as the piece-wise constant function, necessitated the use of a high-resolution grid in order to capture detailed features that demanded vast computational resources. This study proposes an improved fused feature network as well as a comprehensive framework for solving shape classification and segmentation tasks using a two-branch technique and feature learning. We begin by designing a feature encoding network with two distinct building blocks: layer skips within, batch normalization (BN), and rectified linear units (ReLU) in between. The purpose of using layer skips is to have fewer layers to propagate across, which will speed up the learning process and lower the effect of gradients vanishing. Furthermore, we develop a robust grid feature extraction module that consists of multiple convolution blocks accompanied by max-pooling to represent a hierarchical representation and extract features from an input grid. We overcome the grid size constraints by sampling a constant number of points in each grid using a simple K-points nearest neighbor (KNN) search, which aids in learning approximation functions in higher order. The proposed method outperforms or is comparable to state-of-the-art approaches in point cloud segmentation and classification tasks. In addition, a study of ablation is presented to show the effectiveness of the proposed method.
引用
收藏
页数:16
相关论文
共 75 条
  • [1] DPRNet: Deep 3D Point Based Residual Network for Semantic Segmentation and Classification of 3D Point Clouds
    Arshad, Saira
    Shahzad, Muhammad
    Riaz, Qaiser
    Fraz, Muhammad Moazam
    [J]. IEEE ACCESS, 2019, 7 : 68892 - 68904
  • [2] Point Convolutional Neural Networks by Extension Operators
    Atzmon, Matan
    Maron, Haggai
    Lipman, Yaron
    [J]. ACM TRANSACTIONS ON GRAPHICS, 2018, 37 (04):
  • [3] GIFT: A Real-time and Scalable 3D Shape Search Engine
    Bai, Song
    Bai, Xiang
    Zhou, Zhichao
    Zhang, Zhaoxiang
    Latecki, Longin Jan
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 5023 - 5032
  • [4] FFPointNet: Local and global fused feature for 3D point clouds analysis
    Bello, Saifullahi Aminu
    Wang, Cheng
    Wambugu, Naftaly Muriuki
    Adam, Jibril Muhammad
    [J]. NEUROCOMPUTING, 2021, 461 : 55 - 62
  • [5] Review: Deep Learning on 3D Point Clouds
    Bello, Saifullahi Aminu
    Yu, Shangshu
    Wang, Cheng
    Adam, Jibril Muhmmad
    Li, Jonathan
    [J]. REMOTE SENSING, 2020, 12 (11)
  • [6] Pointwise Convolutional Neural Networks
    Binh-Son Hua
    Minh-Khoi Tran
    Yeung, Sai-Kit
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 984 - 993
  • [7] Brock A, 2016, Arxiv, DOI [arXiv:1608.04236, 10.48550/arXiv.1608.04236]
  • [8] Chen Z., 2021, arXiv
  • [9] Chiotellis I., 2016, European Conference on Computer Vision (ECCV), DOI [10.1007/978-3-319-46475-6_21, DOI 10.1007/978-3-319-46475-6_21]
  • [10] 3D-R2N2: A Unified Approach for Single and Multi-view 3D Object Reconstruction
    Choy, Christopher B.
    Xu, Danfei
    Gwak, Jun Young
    Chen, Kevin
    Savarese, Silvio
    [J]. COMPUTER VISION - ECCV 2016, PT VIII, 2016, 9912 : 628 - 644