Deep learning based 3D segmentation in computer vision: A survey

被引:8
作者
He, Yong [1 ]
Yu, Hongshan [1 ]
Liu, Xiaoyan [1 ]
Yang, Zhengeng [2 ]
Sun, Wei [1 ]
Anwar, Saeed [3 ]
Mian, Ajmal [4 ]
机构
[1] Hunan Univ, Quanzhou Inst Ind Design & Machine Intelligence In, Coll Elect & Informat Engn, Sch Robot, Lushan South Rd, Changsha 410082, Hunan, Peoples R China
[2] Hunan Normal Univ, Lushan South Rd, Changsha 410081, Hunan, Peoples R China
[3] Australian Natl Univ, Canberra, ACT 2600, Australia
[4] Univ Western Australia, 35 Stirling Hwy, Perth, WA 6009, Australia
基金
澳大利亚研究理事会; 中国国家自然科学基金;
关键词
Computer vision; Deep learning; Deep neural network; 3D semantic segmentation; 3D instance segmentation; 3D part segmentation; SEMANTIC SEGMENTATION; POINT; NETWORKS; LIDAR;
D O I
10.1016/j.inffus.2024.102722
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
3D segmentation is a fundamental and challenging problem in computer vision with applications in autonomous driving and robotics. It has received significant attention from the computer vision, graphics and machine learning communities. Conventional methods for 3D segmentation, based on hand-crafted features and machine learning classifiers, lack generalization ability. Driven by their success in 2D computer vision, deep learning techniques have recently become the tool of choice for 3D segmentation tasks. This has led to an influx of many methods in the literature that have been evaluated on different benchmark datasets. Whereas survey papers on RGB-D and point cloud segmentation exist, there is a lack of a recent in-depth survey that covers all 3D data modalities and application domains. This paper fills the gap and comprehensively surveys the recent progress in deep learning-based 3D segmentation techniques. We cover over 230 works from the last six years, analyze their strengths and limitations, and discuss their competitive results on benchmark datasets. The survey provides a summary of the most commonly used pipelines and finally highlights promising research directions for the future.
引用
收藏
页数:24
相关论文
共 234 条
[1]   RangeViT: Towards Vision Transformers for 3D Semantic Segmentation in Autonomous Driving [J].
Ando, Angelika ;
Gidaris, Spyros ;
Bursuc, Andrei ;
Puy, Gilles ;
Boulch, Alexandre ;
Marlet, Renaud .
2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, :5240-5250
[2]  
[Anonymous], 2017, P IEEE C COMP VIS PA
[3]   3D Semantic Parsing of Large-Scale Indoor Spaces [J].
Armeni, Iro ;
Sener, Ozan ;
Zamir, Amir R. ;
Jiang, Helen ;
Brilakis, Ioannis ;
Fischer, Martin ;
Savarese, Silvio .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :1534-1543
[4]   An artificial neural network framework for classifying the style of cypriot hybrid examples of built heritage in 3D [J].
Artopoulos, Georgios ;
Maslioukova, Maria I. ;
Zavou, Christina ;
Loizou, Marios ;
Deligiorgi, Marissia ;
Averkiou, Melinos .
JOURNAL OF CULTURAL HERITAGE, 2023, 63 :135-147
[5]  
Atik M.E., 2024, IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens.
[6]   SemanticKITTI: A Dataset for Semantic Scene Understanding of LiDAR Sequences [J].
Behley, Jens ;
Garbade, Martin ;
Milioto, Andres ;
Quenzel, Jan ;
Behnke, Sven ;
Stachniss, Cyrill ;
Gall, Juergen .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :9296-9306
[7]   Review: Deep Learning on 3D Point Clouds [J].
Bello, Saifullahi Aminu ;
Yu, Shangshu ;
Wang, Cheng ;
Adam, Jibril Muhmmad ;
Li, Jonathan .
REMOTE SENSING, 2020, 12 (11)
[8]   Camera-LIDAR Integration: Probabilistic Sensor Fusion for Semantic Mapping [J].
Berrio, Julie Stephany ;
Shan, Mao ;
Worrall, Stewart ;
Nebot, Eduardo .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (07) :7637-7652
[9]   Pointwise Convolutional Neural Networks [J].
Binh-Son Hua ;
Minh-Khoi Tran ;
Yeung, Sai-Kit .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :984-993
[10]   SnapNet: 3D point cloud semantic labeling with 2D deep segmentation networks [J].
Boulch, Alexandre ;
Guerry, Yids ;
Le Saux, Bertrand ;
Audebert, Nicolas .
COMPUTERS & GRAPHICS-UK, 2018, 71 :189-198