ArXiv TLDR

Sovereign Agentic Loops: Decoupling AI Reasoning from Execution in Real-World Systems

🐦 Tweet
2604.22136

Jun He, Deying Yu

cs.CRcs.LG

TLDR

Sovereign Agentic Loops (SAL) decouple AI reasoning from execution, validating LLM intents against system state and policy to enhance safety.

Key contributions

  • Decouples AI reasoning from execution to prevent unsafe actions in real-world systems.
  • Models emit structured intents with justifications, validated against true system state and policy.
  • Combines an obfuscation membrane with a cryptographically linked Evidence Chain for auditability.
  • Formalized to provide policy-bounded execution, identity isolation, and deterministic replay.

Why it matters

As LLM agents increasingly interact with real systems, directly executing stochastic model outputs poses significant safety risks. This paper introduces SAL, a crucial architecture that ensures AI actions are validated against system state and policy, making real-world LLM deployments safer and more auditable. It addresses a critical gap in current agentic systems.

Original Abstract

Large language model (LLM) agents increasingly issue API calls that mutate real systems, yet many current architectures pass stochastic model outputs directly to execution layers. We argue that this coupling creates a safety risk because model correctness, context awareness, and alignment cannot be assumed at execution time. We introduce Sovereign Agentic Loops (SAL), a control-plane architecture in which models emit structured intents with justifications, and the control plane validates those intents against true system state and policy before execution. SAL combines an obfuscation membrane, which limits model access to identity-sensitive state, with a cryptographically linked Evidence Chain for auditability and replay. We formalize SAL and show that, under the stated assumptions, it provides policy-bounded execution, identity isolation, and deterministic replay. In an OpenKedge prototype for cloud infrastructure, SAL blocks 93% of unsafe intents at the policy layer, rejects the remaining 7% via consistency checks, prevents unsafe executions in our benchmark, and adds 12.4 ms median latency.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.