VATLD: A Visual Analytics System to Assess, Understand and Improve Traffic Light Detection

被引:56
作者
Gou, Liang [1 ]
Zou, Lincan [1 ]
Li, Nanxiang [1 ]
Hofmann, Michael [2 ]
Shekar, Arvind Kumar [2 ]
Wendt, Axel [2 ]
Ren, Liu [1 ]
机构
[1] Robert Bosch Res & Technol Ctr, Pittsburgh, PA 15222 USA
[2] Robert Bosch GmbH, Gerlingen, Germany
关键词
Traffic light detection; representation learning; semantic adversarial learning; model diagnosing; autonomous driving;
D O I
10.1109/TVCG.2020.3030350
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Traffic light detection is crucial for environment perception and decision-making in autonomous driving. State-of-the-art detectors are built upon deep Convolutional Neural Networks (CNNs) and have exhibited promising performance. However, one looming concern with CNN based detectors is how to thoroughly evaluate the performance of accuracy and robustness before they can be deployed to autonomous vehicles. In this work, we propose a visual analytics system, VATLD, equipped with a disentangled representation learning and semantic adversarial learning, to assess, understand, and improve the accuracy and robustness of traffic light detectors in autonomous driving applications. The disentangled representation learning extracts data semantics to augment human cognition with human-friendly visual summarization, and the semantic adversarial learning efficiently exposes interpretable robustness risks and enables minimal human interaction for actionable insights. We also demonstrate the effectiveness of various performance improvement strategies derived from actionable insights with our visual analytics system, VATLD, and illustrate some practical implications for safety-critical applications in autonomous driving.
引用
收藏
页码:261 / 271
页数:11
相关论文
共 47 条
[1]   Do Convolutional Neural Networks Learn Class Hierarchy? [J].
Alsallakh, Bilal ;
Jourabloo, Amin ;
Ye, Mao ;
Liu, Xiaoming ;
Ren, Liu .
IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2018, 24 (01) :152-162
[2]   Network Dissection: Quantifying Interpretability of Deep Visual Representations [J].
Bau, David ;
Zhou, Bolei ;
Khosla, Aditya ;
Oliva, Aude ;
Torralba, Antonio .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :3319-3327
[3]  
Behrendt Karsten, 2017, 2017 IEEE International Conference on Robotics and Automation (ICRA), P1370, DOI 10.1109/ICRA.2017.7989163
[4]  
Behrendt K., 2020, GIT REPO BOSCH SMALL
[5]  
Bengio S, 2019, 5 INT C LEARN REPR I
[6]  
Besold T.R., 2017, Neural-symbolic learning and reasoning: A survey and interpretation
[7]  
Burgess Christopher P., 2018, CoRR abs/1804.03599
[8]  
Carroll J. M., 1990, NURNBERG FUNNEL DESI
[9]  
Carter S., 2017, DISTILL, V2, pe9, DOI DOI 10.23915/DISTILL.00009
[10]   ShapeShifter: Robust Physical Adversarial Attack on Faster R-CNN Object Detector [J].
Chen, Shang-Tse ;
Cornelius, Cory ;
Martin, Jason ;
Chau, Duen Horng .
MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2018, PT I, 2019, 11051 :52-68