You can鈥檛 secure what you can鈥檛 categorize: A taxonomy for AI agents
You can鈥檛 secure what you can鈥檛 categorize: A taxonomy for AI agents
Last quarter, the VP of engineering at a large gaming company explained to how they鈥檇 built an AI site reliability engineering agent to help resolve incidents and fix production issues. For weeks, it worked beautifully鈥攖riaging alerts, identifying root causes, even suggesting fixes their team would have taken hours to develop.
Then one day, it launched a denial-of-service attack on their internal monitoring system.
The agent had permissions to query their monitoring APIs. It was supposed to use them to gather context for incident response. But when it decided those APIs might hold the answer to a particularly thorny issue, it started hammering them with requests until the system fell over.
They shut the agent down, obviously. But unplugging the agent is a blunt instrument鈥攊t means losing all the goodness they were getting before.
An agent is a system. To secure any system, you need the right mental model to reason about it. The industry doesn鈥檛 have that mental model for agents yet, and that鈥檚 a problem.
Without a shared mental model of what an agent is, people can鈥檛 decompose it. And if it can鈥檛 be decomposed, security can鈥檛 be designed around it. The disasters . More commonly, though, concerns about agent security are leading to .
Nondeterminism is both the promise and the peril of agents. An AI agent behaves in nondeterministic ways because it is given the agency to determine how it executes tasks. You can鈥檛 remove that autonomy without gutting the agent鈥攂ut you can mitigate the risks. The most fundamental control is permissions.
Consider Wade Foster鈥檚 sharp鈥 and prior work by,, and. If you can categorize what kind of system you鈥檙e building, you can reason about what could go wrong and how to prevent it. Many organizations want to move from left to right on a spectrum of autonomy, but most are stuck because they can鈥檛 reason about what agents might do. Taxonomy can be a diagnostic tool. Know what鈥檚 nondeterministic, and you鈥檒l know where the risk is and what controls to apply.
Agent Taxonomy
Imagine you鈥檙e a retailer. When you get customer feedback, you want to ask happy customers to leave reviews and fix issues for unhappy ones, but you want to automate this. You could build a straightforward automated workflow, or you could move from left to right on this spectrum of autonomy.
Automated Workflow
Automate this as a set of deterministic steps. Store the feedback in the CRM, use a classical ML model to score sentiment, check if it鈥檚 positive or negative, then branch: For positive feedback, send a templated review request with the customer鈥檚 name merged in. For negative feedback, check whether they鈥檙e a small or large customer, then either send a templated apology or create a support ticket with a formulaic summary of their history.
Definition: Deterministic steps or nodes, automated in code or with a workflow automation tool.
What鈥檚 deterministic: Everything
What鈥檚 nondeterministic: Nothing
Security assumptions you can safely make: You know exactly what this system will do.
LLM Workflow
As you move right on the spectrum, replace one or more steps with an LLM鈥攗sually content generation. Now instead of a template apology, an LLM writes a customized response based on the specific feedback. Or it generates a more nuanced summary of customer history for the support team.
Definition: An automated workflow with an LLM used to execute one or more steps.
What鈥檚 deterministic: The steps taken and the control flow between them.
What鈥檚 non-deterministic: Actions taken inside a step (e.g., content generation).
Security assumptions you can safely make: 鈥淚 know what it will do, but not what it will say.鈥
Agentic Workflow
Now you鈥檙e entering agentic territory. An LLM not only produces content but also reasons about control flow. For negative feedback, hand the rest of the process to an agent with access to tools: It can read customer history, send emails, or write to the support queue. The agent decides which tools to use and in what order鈥攎aybe it checks history first, or maybe it sends an immediate apology. You鈥檝e bounded its options, but you haven鈥檛 prescribed the path.
Wade鈥檚 framework defines agentic workflows differently: An LLM is used in multiple steps, but each step remains self-contained and the flow between them is deterministic. That鈥檚 reasonable for demonstrating the value ladder of AI automation. But for security, you need a brighter line. The question is: Does the LLM manage any of the control flow? If it does, you need to reason about all possible paths it might take, not just the content it might generate. That鈥檚 a fundamentally different security posture.
Definition: An automated workflow where part but not all of the control flow is managed by an LLM.
What鈥檚 deterministic: Some control flow.
What鈥檚 non-deterministic: Step content, some control flow.
Security assumptions you can safely make: You know the boundaries of possible paths, but not which path it will take.
Agent
An agent does the whole thing. It gets the raw customer feedback and decides everything: Is it positive or negative? What鈥檚 the customer鈥檚 history? Should the agent apologize, escalate, ask for a review, or something else entirely? It reasons about what tools to use, uses them, and solves the task end-to-end.
Something is a full agent only if it has this end-to-end agency. Any situation where you explicitly lay out the steps doesn鈥檛 qualify鈥攊ncluding workflow automation tools, even when they lean heavily on LLMs. This level of nondeterministic behavior requires a different security posture to respond to all the things an agent could do.
Definition: A task executed end-to-end by an LLM.
What鈥檚 deterministic: Nothing
What鈥檚 non-deterministic: Everything
Security assumptions you can safely make: It will only use tools it can access, but how and whether it will use them is unknown.
Summary
Note on agentic systems: The industry uses 鈥渁gentic systems鈥 as an umbrella term for agentic workflows, agents, and multi-agent systems. From a security perspective, treat every agentic system as equivalent to a full agent, except to the extent that you can point at deterministic controls that bound that agency.
Implications for Securing Agents
You can frame the security implications of agents in different ways, and each one means something different for how you would solve it.
Some say, 鈥淛ust solve prompt injection, and there won鈥檛 be any problems.鈥 Others point to model quality, which is out of your hands (unless you work at a frontier AI lab). Still others frame it as a data loss problem, but data loss has never been solved, even outside AI.
The risk vectors are everywhere鈥攕ee the for a taste. No single framing will capture everything that could go wrong.
Nondeterminism is a feature, not a bug鈥攖hough it comes with security implications. You can鈥檛 remove it without removing the agent鈥檚 agency and demoting it on the spectrum of autonomy.
So don鈥檛 fight nondeterminism. Bound it instead. Play on its home court where it makes sense鈥攅.g., applying agentic oversight to content generation and reasoning. For the really dangerous areas (tool access, data exposure), constrain behaviors with deterministic controls.
What鈥檚 the OG deterministic control for governing who can do what? Permissions.
Permissions for Agents
Permissions are part of the basic infrastructure of any real application. But the state of permissions is not healthy.
Overpermissioning is the status quo. Analysis of Oso permissions data confirms this. What could you鈥攐r an agent with your permissions鈥攄o that would be bad?
One reason people freak out about agents: They intuitively connect these dots. They know people are overpermissioned, they know agents behave nondeterministically, and they can foresee future disasters. 鈥淚 accidentally deleted that Salesforce record once and the system just let me do it. What鈥檚 going to happen if I ask an agent to update Salesforce for me?鈥
If organizations replicate the overpermissioned state of humans in automated systems, what鈥檚 the danger?
- Automated workflow: Low risk because the code does what it鈥檚 programmed to do.
- LLM workflow: Content risk because it might say something wrong or inappropriate.
- Agentic workflow: Action risk because it might take unexpected paths.
- Agent: Maximum action risk because it might do anything it has access to.
An agent should only ever have the permissions for the task at hand. That would mitigate most of the risk. But scoping permissions to match nondeterministic behavior is hard: The agent needs to read customer history and send emails to customers, but you can鈥檛 predict exactly which customers or what it will say. How can you be certain it won鈥檛 leak information?
This taxonomy shows you what you鈥檙e building. It doesn鈥檛 show you how to make it safe.
That gaming company faced a choice between useful and dangerous. The entire industry faces that choice right now. People can build powerful agents or they can build safe agents, but not yet both.
This is supposed to be the decade of agents. But that only happens if people can trust them. That means building infrastructure that doesn鈥檛 exist yet: simulation to test dangerous paths, enforcement that tightens permissions automatically, detection that catches drift, visibility that shows what actually happened.
The taxonomy maps the problem. Now someone needs to build the solution. That鈥檚 the work that matters鈥攏ot because it鈥檚 technically interesting, but because it鈥檚 what unlocks everything else agents could be.
was produced by and reviewed and distributed by 麻豆原创.