Evaluating the Robustness of Uncertainty Quantification-Based Misbehavior Predictors for Autonomous Driving Systems: A Case Study
This program is tentative and subject to change.
Misbehavior predictors play a crucial role in safety-critical autonomous driving systems, as they help anticipate and mitigate risks by signaling potential misbehavior or deviations from expected behavior. Among existing predictors, uncertainty quantification (UQ)-based misbehavior predictors are recognized for their strong capability in identifying these risks and lightweight overhead. These predictors often operate on the assumption that uncertainty signals possible misbehavior, aiding in the real-time recognition of unexpected situations. However, in real-world scenarios, this assumption might not always hold, as many cases are either uncertain but correct or certain but incorrect. This limitation harms the reliability of UQ-based predictors and could lead to safety issues in practice.
In this paper, we conduct a case study to explore the robustness of UQ-based predictors. Here, robustness indicates the ability of predictors to handle two types of challenging test cases, (1) uncertain but correct cases, and (2) certain but incorrect cases. To do so, we develop a generic algorithm-based test data generation framework to produce these two types of test cases and evaluate misbehavior predictors accordingly. Experimental results on MC dropout models with three different settings and seven simulations with various environments demonstrated that even though UQ-based predictors perform well on the original test cases, there is a significant performance degradation when facing challenging test cases. For instance, uncertain but correct cases lead to a 18 F$_3$ score drop to the predictor. These findings highlight critical limitations and emphasize caution when deploying such predictors in unpredictable real-world environments.
This program is tentative and subject to change.
Tue 29 AprDisplayed time zone: Eastern Time (US & Canada) change
11:00 - 12:30 | |||
11:00 18mTalk | AI-Augmented Metamorphic Testing for Comprehensive Validation of Autonomous Vehicles SE4ADS Yueteng Zhang University of Ottawa, Burak Kantarci University of Ottawa, Umair Siddique reasonX Labs Inc. | ||
11:18 18mTalk | Conflict-based Scenario Generation for Autonomous Driving System Validation SE4ADS Hua Qi Kyushu University, Japan, Siyuan Chen The University of Tokyo, Fuyuan Zhang Kyushu University, Tomoyuki TSUCHIYA TIER IV, Michio HAYASHI TIER IV North America, Manabu OKADA TIER IV, Lei Ma The University of Tokyo & University of Alberta, Jianjun Zhao Kyushu University | ||
11:36 18mTalk | Deep Driving Workshop for Education and Training of Behaviour-Based End-to-End Learning Autonomous Driving Systems SE4ADS Mohamed Benchat Institut für Software and Systems Engineering, TU Clausthal, Germany, Iqra Aslam Institut für Software and Systems Engineering, TU Clausthal, Germany, Meng Zhang Institut für Software and Systems Engineering, TU Clausthal, Germany, Nour Habib Institut für Software and Systems Engineering, TU Clausthal, Germany, Abhishek Buragohain Institut für Software and Systems Engineering, TU Clausthal, Germany, Vaibhav Tiwari Institut für Software and Systems Engineering, TU Clausthal, Germany, Andreas Rausch Clausthal University of Technology | ||
11:54 18mTalk | Evaluating the Robustness of Uncertainty Quantification-Based Misbehavior Predictors for Autonomous Driving Systems: A Case Study SE4ADS Xiongfei Wu Kyushu University, Qiang Hu The University of Tokyo, Tomoyuki TSUCHIYA TIER IV, Michio HAYASHI TIER IV North America, Manabu OKADA TIER IV, Lei Ma The University of Tokyo & University of Alberta, Jianjun Zhao Kyushu University | ||
12:12 18mTalk | Moral Testing of Autonomous Driving Systems SE4ADS Wenbing Tang Nanyang Technological University, Mingfei Cheng Singapore Management University, Yuan Zhou Zhejiang Sci-Tech University, Yang Liu Nanyang Technological University |