An AI Agent Execution Environment to Safeguard User Data
Robert Stanley, Avi Verma, Lillian Tsai, Konstantinos Kallas, Sam Kumar
TLDR
GAAP is an AI agent execution environment that guarantees user data confidentiality by enforcing dynamic permissions and tracking data flow, even against attacks.
Key contributions
- Introduces GAAP, an execution environment ensuring user data confidentiality for AI agents.
- Collects dynamic user permissions and enforces them, preventing unauthorized data disclosure.
- Tracks private data flow using augmented Information Flow Control across tasks and time.
- Blocks all data disclosure attacks, including prompt injection, without trusting the agent or model.
Why it matters
AI agents need private data but face significant security risks from attacks and untrusted providers. This paper offers a robust solution, GAAP, that deterministically safeguards user data by enforcing permissions and tracking data flow. It's crucial for building trustworthy and secure personal AI assistants.
Original Abstract
AI agents promise to serve as general-purpose personal assistants for their users, which requires them to have access to private user data (e.g., personal and financial information). This poses a serious risk to security and privacy. Adversaries may attack the AI model (e.g., via prompt injection) to exfiltrate user data. Furthermore, sharing private data with an AI agent requires users to trust a potentially unscrupulous or compromised AI model provider with their private data. This paper presents GAAP (Guaranteed Accounting for Agent Privacy), an execution environment for AI agents that guarantees confidentiality for private user data. Through dynamic and directed user prompts, GAAP collects permission specifications from users describing how their private data may be shared, and GAAP enforces that the agent's disclosures of private user data, including disclosures to the AI model and its provider, comply with these specifications. Crucially, GAAP provides this guarantee deterministically, without trusting the agent with private user data, and without requiring any AI model or the user prompt to be free of attacks. GAAP enforces the user's permission specification by tracking how the AI agent accesses and uses private user data. It augments Information Flow Control with novel persistent data stores and annotations that enable it to track the flow of private information both across execution steps within a single task, and also over multiple tasks separated in time. Our evaluation confirms that GAAP blocks all data disclosure attacks, including those that make other state-of-the-art systems disclose private user data to untrusted parties, without a significant impact on agent utility.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.