4.7 Article

VATLD: A Visual Analytics System to Assess, Understand and Improve Traffic Light Detection

出版社

IEEE COMPUTER SOC
DOI: 10.1109/TVCG.2020.3030350

关键词

Traffic light detection; representation learning; semantic adversarial learning; model diagnosing; autonomous driving

向作者/读者索取更多资源

This research introduces a visual analytics system equipped with disentangled representation learning and semantic adversarial learning to assess and improve the accuracy and robustness of traffic light detectors in autonomous driving applications.
Traffic light detection is crucial for environment perception and decision-making in autonomous driving. State-of-the-art detectors are built upon deep Convolutional Neural Networks (CNNs) and have exhibited promising performance. However, one looming concern with CNN based detectors is how to thoroughly evaluate the performance of accuracy and robustness before they can be deployed to autonomous vehicles. In this work, we propose a visual analytics system, VATLD, equipped with a disentangled representation learning and semantic adversarial learning, to assess, understand, and improve the accuracy and robustness of traffic light detectors in autonomous driving applications. The disentangled representation learning extracts data semantics to augment human cognition with human-friendly visual summarization, and the semantic adversarial learning efficiently exposes interpretable robustness risks and enables minimal human interaction for actionable insights. We also demonstrate the effectiveness of various performance improvement strategies derived from actionable insights with our visual analytics system, VATLD, and illustrate some practical implications for safety-critical applications in autonomous driving.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据