all AI news
Predicting Safety Misbehaviours in Autonomous Driving Systems using Uncertainty Quantification
April 30, 2024, 4:42 a.m. | Ruben Grewal, Paolo Tonella, Andrea Stocco
cs.LG updates on arXiv.org arxiv.org
Abstract: The automated real-time recognition of unexpected situations plays a crucial role in the safety of autonomous vehicles, especially in unsupported and unpredictable scenarios. This paper evaluates different Bayesian uncertainty quantification methods from the deep learning domain for the anticipatory testing of safety-critical misbehaviours during system-level simulation-based testing. Specifically, we compute uncertainty scores as the vehicle executes, following the intuition that high uncertainty scores are indicative of unsupported runtime conditions that can be used to distinguish …
abstract arxiv automated autonomous autonomous driving autonomous driving systems autonomous vehicles bayesian cs.lg cs.se deep learning domain driving paper quantification real-time recognition role safety safety-critical systems testing type uncertainty vehicles
More from arxiv.org / cs.LG updates on arXiv.org
Testing the Segment Anything Model on radiology data
1 day, 11 hours ago |
arxiv.org
Calorimeter shower superresolution
1 day, 11 hours ago |
arxiv.org
Jobs in AI, ML, Big Data
Software Engineer for AI Training Data (School Specific)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Python)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Tier 2)
@ G2i Inc | Remote
Data Engineer
@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US