SMCNet: State-Space Model for Enhanced Corruption Robustness in 3D Classification

被引:0
作者
Li, Junhui [1 ]
Huang, Bangju [1 ]
Pan, Lei [2 ]
机构
[1] Civil Aviat Flight Univ China, Coll Air Traff Management, Deyang 618307, Peoples R China
[2] Civil Aviat Flight Univ China, Sch Comp Sci, Deyang 618307, Peoples R China
关键词
point cloud; state-space model; object classification; LiDAR; corruption robustness; NETWORK;
D O I
10.3390/s24237861
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
Accurate classification of three-dimensional (3D) point clouds in real-world environments is often impeded by sensor noise, occlusions, and incomplete data. To overcome these challenges, we propose SMCNet, a robust multimodal framework for 3D point cloud classification. SMCNet combines multi-view projection and neural radiance fields (NeRFs) to generate high-fidelity 2D representations with enhanced texture realism, addressing occlusions and lighting inconsistencies effectively. The Mamba model is further refined within this framework by integrating a depth perception module to capture long-range point interactions and adopting a dual-channel structure to enhance point-wise feature extraction. Fine-tuning adapters for the CLIP and Mamba models are also introduced, significantly improving cross-domain adaptability. Additionally, an intelligent voting mechanism aggregates predictions from multiple viewpoints, ensuring enhanced classification robustness. Comprehensive experiments demonstrate that SMCNet achieves state-of-the-art performance, outperforming the PointNet++ baseline with a 0.5% improvement in mean overall accuracy (mOA) on ModelNet40 and a 7.9% improvement on ScanObjectNN. In corruption resistance, SMCNet reduces the mean corruption error (mCE) by 0.8% on ModelNet40-C and 3.6% on ScanObjectNN-C. These results highlight the effectiveness of SMCNet in tackling real-world classification scenarios with noisy and corrupted data.
引用
收藏
页数:20
相关论文
共 45 条
  • [1] MS-TCN: Multi-Stage Temporal Convolutional Network for Action Segmentation
    Abu Farha, Yazan
    Gall, Juergen
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 3570 - 3579
  • [2] DGCNN: A convolutional neural network over large-scale labeled graphs
    Anh Viet Phan
    Minh Le Nguyen
    Yen Lam Hoang Nguyen
    Lam Thu Bui
    [J]. NEURAL NETWORKS, 2018, 108 : 533 - 543
  • [3] Points to Patches: Enabling the Use of Self-Attention for 3D Shape Recognition
    Berg, Axel
    Oskarsson, Magnus
    O'Connor, Mark
    [J]. 2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 528 - 534
  • [4] PRA-Net: Point Relation-Aware Network for 3D Point Cloud Analysis
    Cheng, Silin
    Chen, Xiwu
    He, Xinwei
    Liu, Zhe
    Bai, Xiang
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 : 4436 - 4448
  • [5] Goyal A, 2021, PR MACH LEARN RES, V139
  • [6] RSCNN: A CNN-Based Method to Enhance Low-Light Remote-Sensing Images
    Hu, Linshu
    Qin, Mengjiao
    Zhang, Feng
    Du, Zhenhong
    Liu, Renyi
    [J]. REMOTE SENSING, 2021, 13 (01) : 1 - 13
  • [7] Huang RZ, 2020, PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P2355
  • [8] Point Cloud Augmentation with Weighted Local Transformations
    Kim, Sihyeon
    Lee, Sanghyeok
    Hwang, Dasol
    Lee, Jaewon
    Hwang, Seong Jae
    Kim, Hyunwoo J.
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 528 - 537
  • [9] Regularization Strategy for Point Cloud via Rigidly Mixed Sample
    Lee, Dogyoon
    Lee, Jaeha
    Lee, Junhyeop
    Lee, Hyeongmin
    Lee, Minhyeok
    Woo, Sungmin
    Lee, Sangyoun
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 15895 - 15904
  • [10] Li J., 2021, IEEE Trans. Robot, V37, P1123