Fail-Safe Human Detection for Drones Using a Multi-Modal Curriculum Learning Approach

被引:9
作者
Safa, Ali [1 ,2 ]
Verbelen, Tim [3 ,4 ]
Ocket, Ilja [1 ,2 ]
Bourdoux, Andre [1 ]
Catthoor, Francky [1 ,2 ]
Gielen, Georges G. E. [1 ,2 ]
机构
[1] IMEC, B-3001 Leuven, Belgium
[2] Katholieke Univ Leuven, B-3001 Leuven, Belgium
[3] Univ Ghent, IMEC, B-9000 Ghent, Belgium
[4] Univ Ghent, IDLab, B-9000 Ghent, Belgium
关键词
Radar; Cameras; Drones; Radar imaging; Voltage control; Standards; Imaging; People detection; sensor fusion; UAVs; curriculum learning; AUTOMOTIVE RADAR;
D O I
10.1109/LRA.2021.3125450
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Drones are currently being explored for safety-critical applications where human agents are expected to evolve in their vicinity. In such applications, robust people avoidance must be provided by fusing a number of sensing modalities in order to avoid collisions. Currently however, people detection systems used on drones are solely based on standard cameras besides an emerging number of works discussing the fusion of imaging and event-based cameras. On the other hand, radar-based systems provide up-most robustness towards environmental conditions but do not provide complete information on their own and have mainly been investigated in automotive contexts, not for drones. In order to enable the fusion of radars with both event-based and standard cameras, we present KUL-UAVSAFE, a first-of-its-kind dataset for the study of safety-critical people detection by drones. In addition, we propose a baseline CNN architecture with cross-fusion highways and introduce a curriculum learning strategy for multi-modal data termed SAUL, which greatly enhances the robustness of the system towards hard RGB failures and provides a significant gain of 15% in peak F-1 score compared to the use of BlackIn, previously proposed for cross-fusion networks. We demonstrate the real-time performance and feasibility of the approach by implementing the system in an edge-computing unit. We release our dataset and additional material in the project home page.
引用
收藏
页码:303 / 310
页数:8
相关论文
共 27 条
[1]  
[Anonymous], SQUEEZENET ALEXNET L
[2]   Radar-camera Fusion for Road Target Classification [J].
Aziz, Kheireddine ;
De Greef, Eddy ;
Rykunov, Maxim ;
Bourdoux, Andre ;
Sahli, Hichem .
2020 IEEE RADAR CONFERENCE (RADARCONF20), 2020,
[3]  
Bengio Y., 2009, P 26 ANN INT C MACHI, P41
[4]  
Bochkovskiy A., 2020, ARXIV 200410934
[5]   A 240 x 180 130 dB 3 μs Latency Global Shutter Spatiotemporal Vision Sensor [J].
Brandli, Christian ;
Berner, Raphael ;
Yang, Minhao ;
Liu, Shih-Chii ;
Delbruck, Tobi .
IEEE JOURNAL OF SOLID-STATE CIRCUITS, 2014, 49 (10) :2333-2341
[6]  
Caesar H, 2020, PROC CVPR IEEE, P11618, DOI 10.1109/CVPR42600.2020.01164
[7]  
Delmerico J, 2019, IEEE INT CONF ROBOT, P6713, DOI [10.1109/icra.2019.8793887, 10.1109/ICRA.2019.8793887]
[8]   Tinier-YOLO: A Real-Time Object Detection Method for Constrained Environments [J].
Fang, Wei ;
Wang, Lin ;
Ren, Peiming .
IEEE ACCESS, 2020, 8 :1935-1944
[9]   Event-Based Vision: A Survey [J].
Gallego, Guillermo ;
Delbruck, Tobi ;
Orchard, Garrick Michael ;
Bartolozzi, Chiara ;
Taba, Brian ;
Censi, Andrea ;
Leutenegger, Stefan ;
Davison, Andrew ;
Conradt, Jorg ;
Daniilidis, Kostas ;
Scaramuzza, Davide .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (01) :154-180
[10]   RAMP-CNN: A Novel Neural Network for Enhanced Automotive Radar Object Recognition [J].
Gao, Xiangyu ;
Xing, Guanbin ;
Roy, Sumit ;
Liu, Hui .
IEEE SENSORS JOURNAL, 2021, 21 (04) :5119-5132