Safe Control using Learned Safety Filters and Adaptive Conformal Inference
Sacha Huriot, Ihab Tabbara, Hussein Sibai
TLDR
ACoFi enhances learned safety filters with adaptive conformal inference, dynamically adjusting switching to provide reliable control and soft safety guarantees.
Key contributions
- Introduces ACoFi, combining learned safety filters with adaptive conformal inference.
- Dynamically adjusts policy switching based on observed prediction errors to enhance safety.
- Quantifies uncertainty in nominal policy safety, providing soft safety guarantees.
- Outperforms baselines, reducing safety violations in out-of-distribution scenarios.
Why it matters
This paper addresses a critical challenge in learned safety filters: their reliability due to prediction errors. ACoFi offers a novel solution by dynamically adapting switching criteria and quantifying uncertainty, providing more robust safety guarantees than previous methods. This is crucial for deploying AI in safety-critical systems.
Original Abstract
Safety filters have been shown to be effective tools to ensure the safety of control systems with unsafe nominal policies. To address scalability challenges in traditional synthesis methods, learning-based approaches have been proposed for designing safety filters for systems with high-dimensional state and control spaces. However, the inevitable errors in the decisions of these models raise concerns about their reliability and the safety guarantees they offer. This paper presents Adaptive Conformal Filtering (ACoFi), a method that combines learned Hamilton-Jacobi reachability-based safety filters with adaptive conformal inference. Under ACoFi, the filter dynamically adjusts its switching criteria based on the observed errors in its predictions of the safety of actions. The range of possible safety values of the nominal policy's output is used to quantify uncertainty in safety assessment. The filter switches from the nominal policy to the learned safe one when that range suggests it might be unsafe. We show that ACoFi guarantees that the rate of incorrectly quantifying uncertainty in the predicted safety of the nominal policy is asymptotically upper bounded by a user-defined parameter. This gives a soft safety guarantee rather than a hard safety guarantee. We evaluate ACoFi in a Dubins car simulation and a Safety Gymnasium environment, empirically demonstrating that it significantly outperforms the baseline method that uses a fixed switching threshold by achieving higher learned safety values and fewer safety violations, especially in out-of-distribution scenarios.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.