AziNorm: Exploiting the Radial Symmetry of Point Cloud for Azimuth-Normalized 3D Perception

被引:5
作者
Chen, Shaoyu [1 ,2 ]
Wang, Xinggang [1 ]
Cheng, Tianheng [1 ,2 ]
Zhang, Wenqiang [1 ]
Zhang, Qian [2 ]
Huang, Chang [2 ]
Liu, Wenyu [1 ]
机构
[1] Huazhong Univ Sci & Technol, Sch EIC, Wuhan, Peoples R China
[2] Horizon Robot, Beijing, Peoples R China
来源
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022) | 2022年
关键词
D O I
10.1109/CVPR52688.2022.00628
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Studying the inherent symmetry of data is of great importance in machine learning. Point cloud, the most important data format for 3D environmental perception, is naturally endowed with strong radial symmetry. In this work, we exploit this radial symmetry via a divide-and-conquer strategy to boost 3D perception performance and ease optimization. We propose Azimuth Normalization (AziNorm), which normalizes the point clouds along the radial direction and eliminates the variability brought by the difference of azimuth. AziNorm can be flexibly incorporated into most LiDAR-based perception methods. To validate its effectiveness and generalization ability, we apply AziNorm in both object detection and semantic segmentation. For detection, we integrate AziNorm into two representative detection methods, the one-stage SECOND detector and the state-of-the-art two-stage PV-RCNN detector. Experiments on Waymo Open Dataset demonstrate that AziNorm improves SECOND and PV-RCNN by 7.03 mAPH and 3.01 mAPH respectively. For segmentation, we integrate AziNorm into KPConv. On SemanticKitti dataset, AziNorm improves KPConv by 1.6/1.1 mIoU on val/test set. Besides, AziNorm remarkably improves data efficiency and accelerates convergence, reducing the requirement of data amounts or training epochs by an order of magnitude. SECOND w/ AziNorm can significantly outperform fully trained vanilla SECOND, even trained with only 10% data or 10% epochs. Code and models are available at https://github.com/hustvl/AziNorm.
引用
收藏
页码:6377 / 6386
页数:10
相关论文
共 52 条
[1]  
[Anonymous], 2020, CVPR, DOI DOI 10.1109/CVPR42600.2020.00466
[2]  
[Anonymous], ICCV
[3]  
[Anonymous], 2019, CVPR, DOI DOI 10.1109/CVPR.2019.01298
[4]  
[Anonymous], 2019, NEURIPS
[5]  
[Anonymous], 2020, CVPR, DOI DOI 10.1109/CVPR42600.2020.00033
[6]  
[Anonymous], 2020, CVPR, DOI DOI 10.1109/CVPR42600.2020.00570
[7]  
[Anonymous], 2018, NeurIPS
[8]  
[Anonymous], 2019, CVPR, DOI DOI 10.1109/CVPR.2019.00319
[9]  
[Anonymous], 2019, ICCV, DOI DOI 10.1109/ICCV.2019.01053
[10]   SemanticKITTI: A Dataset for Semantic Scene Understanding of LiDAR Sequences [J].
Behley, Jens ;
Garbade, Martin ;
Milioto, Andres ;
Quenzel, Jan ;
Behnke, Sven ;
Stachniss, Cyrill ;
Gall, Juergen .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :9296-9306