COMMENTARY: A new class of security failure has already taken hold as AI agents push deeper into production. Security blogger Simon Willison's post earlier this year on the “lethal trifecta” crystallized this growing concern.
Willison said an AI agent becomes dangerous when three conditions coexist: untrusted human input, access to sensitive data, and a path to exfiltrate information.
[SC Media Perspectives columns are written by a trusted community of SC Media cybersecurity subject matter experts. Read more Perspectives here.]
If an attacker can trigger all three, they can manipulate agents into stealing private data on command.
Many teams still think of agents as upgraded chatbots that follow explicit user instructions, not as autonomous and non-deterministic entities that can act on their own.
But as AI agents move into production, this trifecta becomes a structural risk that traditional controls were never designed to handle. Here’s where authorization can act as a defining control layer, capable of breaking the trifecta before it becomes an attack.
Moving forward, keep these three points in mind:
The lethal trifecta represents a structural risk that demands deterministic, fine-grained controls enforced outside the model itself. LLMs are probabilistic and manipulable by design. This makes them powerful, but also fundamentally unsuitable as their own security guardians.
Any attempt to use LLMs to reliably detect or block manipulation inherits the same non-determinism and failure modes as the system it’s meant to protect. And that's not good enough for security, where Willison notes that 99% is a failing grade. Authorization offers that deterministic external control layer through:
Task-bound access: Agents should receive only the narrowly scoped permissions required for the task they are executing, for example, read access to a single record rather than an entire table.
Total visibility: Every agent action must be observable. When an agent behaves unexpectedly, teams need an immediate signal.
Instant containment: When behavior goes sideways, teams must quarantine an agent in real time.
Centralized control: Permissions must live in one auditable, testable place. Fragmented configuration across agents and tools guarantees drift and inconsistent enforcement.
The autonomy, speed, and non-deterministic behavior of agents make them both powerful and dangerous in ways the industry has only begun to grasp. The lethal trifecta reminds us that attackers don’t need to compromise our systems; they only need to convince the agent.
Managing the lethal trifecta requires engineering and security working in lockstep. Authorization gives them the shared control point to do it.
Gabe Jackson, founding engineer, Oso
SC Media Perspectives columns are written by a trusted community of SC Media cybersecurity subject matter experts. Each contribution has a goal of bringing a unique voice to important cybersecurity topics. Content strives to be of the highest quality, objective and non-commercial.