RL-STPA: Adapting System-Theoretic Hazard Analysis for Safety-Critical Reinforcement Learning
Steven A. Senczyszyn, Timothy C. Havens, Nathaniel Rice, Jason E. Summers, Benjamin D. Werner + 1 more
TLDR
RL-STPA adapts system-theoretic hazard analysis for safety-critical reinforcement learning by decomposing tasks, using perturbation testing, and iterative feedback.
Key contributions
- Hierarchical subtask decomposition using temporal phase analysis and domain expertise for emergent behaviors.
- Coverage-guided perturbation testing to explore state-action space sensitivity and identify hazards.
- Iterative checkpoints feeding identified hazards back into training via reward shaping and curriculum design.
Why it matters
RL deployments in safety-critical domains lack systematic hazard identification. RL-STPA offers a practical framework to evaluate and improve RL safety and robustness, addressing limitations of existing methods.
Original Abstract
As reinforcement learning (RL) deployments expand into safety-critical domains, existing evaluation methods fail to systematically identify hazards arising from the black-box nature of neural network enabled policies and distributional shift between training and deployment. This paper introduces Reinforcement Learning System-Theoretic Process Analysis (RL-STPA), a framework that adapts conventional STPA's systematic hazard analysis to address RL's unique challenges through three key contributions: hierarchical subtask decomposition using both temporal phase analysis and domain expertise to capture emergent behaviors, coverage-guided perturbation testing that explores the sensitivity of state-action spaces, and iterative checkpoints that feed identified hazards back into training through reward shaping and curriculum design. We demonstrate RL-STPA in the safety-critical test case of autonomous drone navigation and landing, revealing potential loss scenarios that can be missed by standard RL evaluations. The proposed framework provides practitioners with a toolkit for systematic hazard analysis, quantitative metrics for safety coverage assessment, and actionable guidelines for establishing operational safety bounds. While RL-STPA cannot provide formal guarantees for arbitrary neural policies, it offers a practical methodology for systematically evaluating and improving RL safety and robustness in safety-critical applications where exhaustive verification methods remain intractable.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.