Interval POMDP Shielding for Imperfect-Perception Agents
TLDR
This paper introduces an Interval POMDP shielding method to enhance safety in autonomous systems with imperfect perception by blocking unsafe actions.
Key contributions
- Models systems with imperfect perception and known dynamics as Interval POMDPs using confidence intervals.
- Proposes an algorithm to compute a conservative set of beliefs over the underlying system state.
- Develops a runtime shield with a finite-horizon safety guarantee based on learned perception intervals.
- Demonstrates improved safety over state-of-the-art baselines in four experimental case studies.
Why it matters
Autonomous systems with learned perception often make unsafe decisions due to sensor errors. This work provides a robust method to ensure safety by proactively blocking potentially dangerous actions, offering strong guarantees even with uncertain perception models. It's crucial for reliable AI deployment.
Original Abstract
Autonomous systems that rely on learned perception can make unsafe decisions when sensor readings are misclassified. We study shielding for this setting: given a proposed action, a shield blocks actions that could violate safety. We consider the common case where system dynamics are known but perception uncertainty must be estimated from finite labeled data. From these data we build confidence intervals for the probabilities of perception outcomes and use them to model the system as a finite Interval Partially Observable Markov Decision Process with discrete states and actions. We then propose an algorithm to compute a conservative set of beliefs over the underlying state that is consistent with the observations seen so far. This enables us to construct a runtime shield that comes with a finite-horizon guarantee: with high probability over the training data, if the true perception uncertainty rates lie within the learned intervals, then every action admitted by the shield satisfies a stated lower bound on safety. Experiments on four case studies show that our shielding approach (and variants derived from it) improves the safety of the system over state-of-the-art baselines.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.