3D-VField: Adversarial Augmentation of Point Clouds for Domain Generalization in 3D Object Detection

被引:48
作者
Lehner, Alexander [1 ,2 ]
Gasperini, Stefano [1 ,2 ]
Marcos-Ramiro, Alvaro [2 ]
Schmidt, Michael [2 ]
Mahani, Mohammad-Ali Nikouei [2 ]
Navab, Nassir [1 ,3 ]
Busam, Benjamin [1 ]
Tombari, Federico [1 ,4 ]
机构
[1] Tech Univ Munich, Munich, Germany
[2] BMW Grp, Munich, Germany
[3] Johns Hopkins Univ, Baltimore, MD 21218 USA
[4] Google, Mountain View, CA 94043 USA
来源
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022) | 2022年
关键词
D O I
10.1109/CVPR52688.2022.01678
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
As 3D object detection on point clouds relies on the geometrical relationships between the points, non-standard object shapes can hinder a method's detection capability. However, in safety-critical settings, robustness to out-of-domain and long-tail samples is fundamental to circumvent dangerous issues, such as the misdetection of damaged or rare cars. In this work, we substantially improve the generalization of 3D object detectors to out-of-domain data by deforming point clouds during training. We achieve this with 3D-VField: a novel data augmentation method that plausibly deforms objects via vector fields learned in an adversarial fashion. Our approach constrains 3D points to slide along their sensor view rays while neither adding nor removing any of them. The obtained vectors are transferable, sample-independent and preserve shape and occlusions. Despite training only on a standard dataset, such as KITH, augmenting with our vector fields significantly improves the generalization to differently shaped objects and scenes. Towards this end, we propose and share CrashD: a synthetic dataset of realistic damaged and rare cars, with a variety of crash scenarios. Extensive experiments on KITTI, Waymo, our CrashD and SUN RGB-D show the generalizability of our techniques to out-of-domain data, different models and sensors, namely LiDAR and ToF cameras, for both indoor and outdoor scenes. Our CrashD dataset is available at https://crashd-cars.github.io.
引用
收藏
页码:17274 / 17283
页数:10
相关论文
共 45 条
[1]  
Alaifari Rima, 2019, P INT C LEARN REPR
[2]  
Albuquerque I., 2020, ARXIV200313525
[3]  
Balaji Yogesh, 2018, Advances in neural information processing systems, V31, P2
[4]  
Beery S, 2020, IEEE WINT CONF APPL, P852, DOI [10.1109/WACV45572.2020.9093570, 10.1109/wacv45572.2020.9093570]
[5]   Description of Corner Cases in Automated Driving: Goals and Challenges [J].
Bogdoll, Daniel ;
Breitenstein, Jasmin ;
Heidecker, Florian ;
Bieshaar, Maarten ;
Sick, Bernhard ;
Fingscheidt, Tim ;
Zoellner, J. Marius .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW 2021), 2021, :1023-1028
[6]   Unifying Knowledge Graph Learning and Recommendation: Towards a Better Understanding of User Preferences [J].
Cao, Yixin ;
Wang, Xiang ;
He, Xiangnan ;
Hu, Zikun ;
Chua, Tat-Seng .
WEB CONFERENCE 2019: PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE (WWW 2019), 2019, :151-161
[7]   An Efficient and Flexible Learning Framework for Dynamic Power and Thermal Co-Management [J].
Cao, Yuan ;
Shen, Tianhao ;
Zhang, Li ;
Yin, Xunzhao ;
Zhuo, Cheng .
PROCEEDINGS OF THE 2020 ACM/IEEE 2ND WORKSHOP ON MACHINE LEARNING FOR CAD (MLCAD '20), 2020, :117-122
[8]   Adversarial Sensor Attack on LiDAR-based Perception in Autonomous Driving [J].
Cao, Yulong ;
Xiao, Chaowei ;
Cyr, Benjamin ;
Zhou, Yimeng ;
Park, Won ;
Rampazzi, Sara ;
Chen, Qi Alfred ;
Fu, Kevin ;
Mao, Z. Morley .
PROCEEDINGS OF THE 2019 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY (CCS'19), 2019, :2267-2281
[9]   Towards Evaluating the Robustness of Neural Networks [J].
Carlini, Nicholas ;
Wagner, David .
2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2017, :39-57
[10]   R4Dyn: Exploring Radar for Self-Supervised Monocular Depth Estimation of Dynamic Scenes [J].
Gasperini, Stefano ;
Koch, Patrick ;
Dallabetta, Vinzenz ;
Navab, Nassir ;
Busam, Benjamin ;
Tombari, Federico .
2021 INTERNATIONAL CONFERENCE ON 3D VISION (3DV 2021), 2021, :751-760