Safe Continual Reinforcement Learning in Non-stationary Environments
Austin Coursey, Abel Diaz-Gonzalez, Marcos Quinones-Grueiro, Gautam Biswas
TLDR
Explores Safe Continual RL in non-stationary environments, revealing a fundamental trade-off between safety and catastrophic forgetting.
Key contributions
- Introduces three benchmark environments for safety-critical continual adaptation.
- Evaluates representative Safe RL, Continual RL, and combined approaches.
- Identifies a tension between maintaining safety and preventing catastrophic forgetting.
- Examines regularization-based strategies to partially mitigate this safety-forgetting trade-off.
Why it matters
Real-world RL systems must adapt to changing conditions while ensuring safety. This paper highlights critical limitations of current methods in balancing these objectives, guiding future research toward resilient autonomous controllers.
Original Abstract
Reinforcement learning (RL) offers a compelling data-driven paradigm for synthesizing controllers for complex systems when accurate physical models are unavailable; however, most existing control-oriented RL methods assume stationarity and, therefore, struggle in real-world non-stationary deployments where system dynamics and operating conditions can change unexpectedly. Moreover, RL controllers acting in physical environments must satisfy safety constraints throughout their learning and execution phases, rendering transient violations during adaptation unacceptable. Although continual RL and safe RL have each addressed non-stationarity and safety, respectively, their intersection remains comparatively unexplored, motivating the study of safe continual RL algorithms that can adapt over the system's lifetime while preserving safety. In this work, we systematically investigate safe continual reinforcement learning by introducing three benchmark environments that capture safety-critical continual adaptation and by evaluating representative approaches from safe RL, continual RL, and their combinations. Our empirical results reveal a fundamental tension between maintaining safety constraints and preventing catastrophic forgetting under non-stationary dynamics, with existing methods generally failing to achieve both objectives simultaneously. To address this shortcoming, we examine regularization-based strategies that partially mitigate this trade-off and characterize their benefits and limitations. Finally, we outline key open challenges and research directions toward developing safe, resilient learning-based controllers capable of sustained autonomous operation in changing environments.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.