A group of cybersecurity researchers including Khoury faculty Alina Oprea and Cristina Nita-Rotaru pose for a photo in a lab next to a Cybersecurity and Privacy Institute banner

Security of LLM Agents

Postdoctoral researcher

Students

Project description

AI agents with increased levels of autonomy are being deployed in safety critical applications, such as healthcare, finance, and cybersecurity. These agents, built on top of large language models (LLMs), are becoming increasingly competitive at automating complex tasks performed by humans, by leveraging access to tools installed on user devices, external resources, and interacting with other AI agents autonomously. These agents might get access to personal sensitive information, such as email, calendar, photos, and location, especially when running on user devices. Unfortunately, the enhanced functionality of AI agents broadens the attack surface of agentic systems, which includes multiple security risks: prompt injection, agent hijacking, propagating malware, and leakage of sensitive user data. Several challenges in securing AI agentic systems remain, including providing secure mechanisms for agent discovery and communication, agent oversight, and protection against attacks such as prompt injection.

a diagram showing a model of security for LLM agents. There are devices separated into groups of on-device data, LLM agent, external data, and malicious agent.

SAGA: A Security Architecture for Governing AI Agentic Systems. We propose SAGA, a framework for governing LLM agent deployment, designed to enhance security while offering user oversight on their agents’ lifecycle. SAGA employs a centralized Provider  that maintains agents’ contact information and user-defined access control policies, enabling users to register their agents and enforce their policies on inter-agent communication. SAGA introduces a cryptographic mechanism for deriving access control tokens, that offers fine-grained control over an agent’s interaction with other agents, providing formal security guarantees. We evaluate SAGA on several agentic tasks, using agents in different geolocations, and multiple on-device and cloud LLMs. SAGA scales well and can support millions of agents while maintaining minimal overhead throughout the entire agent lifecycle, from registration to deactivation. 

A Security Architecture for LLM-Integrated App Systems. LLM-integrated app systems extend the utility of Large Language Models (LLMs) with third-party apps that are invoked by a system LLM using interleaved planning and execution phases to answer user queries. These systems introduce new attack vectors where malicious apps can cause integrity violations of planning or execution, availability breakdown, or privacy compromise during execution. We propose Abstract-Concrete-Execute (ACE), a new secure architecture for LLM-integrated app systems that provides security guarantees for system planning and execution. Specifically, ACE decouples planning into two phases by first creating an abstract execution plan using only trusted information, and then mapping the abstract plan to a concrete plan using installed system apps. We verify that the plans generated by our system satisfy user-specified secure information flow constraints via static analysis on the structured plan output. During execution, ACE enforces data and capability barriers between apps, and ensures that the execution is conducted according to the trusted abstract plan. Our architecture represents a significant advancement towards hardening LLM-based systems using system security principles.

Relevant publications