Zero Trust Security Faces Its AI Agent Test
AI agents that can buy things and spawn sub-agents need security frameworks that assume breach from the start. Zero trust principles are getting a second life.
Written by AI. Bob Reynolds
February 10, 2026

Photo: IBM Technology / YouTube
Zero trust became a marketing cliché so quickly that the useful ideas underneath nearly suffocated. Every vendor selling anything remotely security-related slapped the label on their pitch deck between 2018 and 2022. But now we have AI agents that can call APIs, move money, and spawn their own sub-agents, and suddenly those buried principles matter again.
Jeff Crume, a cybersecurity architect at IBM, walks through what happens when you apply zero trust thinking to autonomous AI systems in a new video breaking down the security challenges. His framing is useful because he doesn't pretend zero trust is new or revolutionary. He acknowledges the hype cycle exhausted the term. What makes his explanation worth attention is the specificity: here's exactly where traditional security models break when software starts acting independently.
The Attack Surface Just Multiplied
Traditional zero trust concerns itself with users, devices, networks, and data. You verify identity. You encrypt sensitive information. You segment networks so a breach in one area doesn't cascade everywhere. You assume someone has already broken in and design accordingly.
Agentic AI systems require all of that plus several new layers. "The actors are in fact software," Crume notes. "An agent may in fact use lots of these different non-human identities. So here we have a proliferation of these things growing."
Each identity represents a potential entry point. Each API call the agent makes opens another attack vector. Each tool the agent can access extends the surface area an attacker might exploit. And because these systems operate autonomously, they can execute hundreds of actions before a human notices something has gone wrong.
The threat map Crume outlines is extensive. Attackers can inject malicious prompts to hijack the agent's behavior. They can poison the training data or policy preferences that guide decision-making. They can intercept API calls. They can steal credentials. They can compromise any of the tools or databases the agent touches. "An attacker has a wealth of different places that they could in fact dive into and do a lot of damage," he says.
This isn't theoretical paranoia. It's a straightforward accounting of what happens when you give software the ability to act in the world without requiring human approval for each action.
Just-in-Time Credentials and Tool Registries
Crume's proposed defenses follow a pattern: treat AI agents like you should have been treating human users all along, but with less tolerance for convenience shortcuts.
First: unique, dynamic credentials for every agent and every sub-agent those agents create. No more embedding API keys directly in code, something programmers have done for decades despite knowing better. Everything goes in a vault. Credentials get checked out when needed and returned when done. This is the "just in time" principle replacing "just in case" access.
Second: a tool registry functioning as a whitelist. The agent can only interact with APIs, databases, and services that have been vetted and registered as secure. If you're making soup, Crume says, you verify the ingredients are pure before cooking. The metaphor is simple because the concept is simple. The difficulty lies in implementation—maintaining that registry as tools change and multiply.
Third: an AI firewall or gateway that inspects inputs and outputs in real time. This layer watches for prompt injection attempts, checks for data leaking outside approved boundaries, and blocks improper calls before they execute. It's pervasive monitoring rather than perimeter defense.
Fourth: immutable logs. Everything the agent does gets recorded in a way that can't be altered after the fact. "When actions are occurring in the system, it needs to be able to be traceable so we can go back later and understand why it did what it did," Crume explains. This matters both for security forensics and for understanding autonomous decision-making.
The Human Override Question
Crume includes something often missing from AI security discussions: the kill switch. Even with all these controls in place, humans need the ability to shut things down when an agent starts behaving unexpectedly. He also suggests throttles—if you built a purchasing agent, you probably don't want it deciding to buy a thousand units of something in sixty seconds, even if technically that's within its mandate.
These guardrails acknowledge an uncomfortable reality: we're deploying systems whose behavior we can't fully predict. The controls aren't about preventing all possible failures. They're about maintaining the ability to intervene when failures occur.
What's less clear from Crume's framework is who decides when intervention is necessary. In a traditional security breach, the indicators are often obvious—unauthorized access, data exfiltration, service disruption. With an AI agent, the line between "working as intended" and "doing something problematic" can be ambiguous. An agent that finds a creative but questionable solution to a problem might look like either innovation or malfunction depending on perspective.
The video also doesn't address the computational and operational cost of implementing these controls. Real-time inspection of every agent action, immutable logging of every decision, dynamic credential management for potentially thousands of non-human identities—this isn't trivial infrastructure. Small organizations deploying AI agents might find the security overhead as challenging as building the agents themselves.
Beyond the Hype Cycle
What makes Crume's presentation valuable is his willingness to acknowledge that zero trust became marketing noise while maintaining that the core principles remain sound. "I never got confused by all that noise because I knew there were some solid, even game-changing security principles worth holding on to," he says.
The most important of those principles: assume breach. Design your security assuming attackers are already inside your systems, already have elevated privileges, already compromised something. This mindset shift matters more with AI agents because the speed and autonomy of these systems mean a successful attack can accomplish significantly more before detection.
Crume's framing suggests we're at an inflection point similar to when organizations first connected internal networks to the internet or when mobile devices started accessing corporate systems. Each transition required rethinking security models. Each time, the initial response was trying to extend old frameworks to new contexts. Eventually, fundamental redesigns became necessary.
AI agents that can spawn other agents, that can access tools and APIs autonomously, that can move data and execute transactions—these aren't incremental changes to existing systems. They're a different category of capability with a correspondingly different threat profile. Whether zero trust principles as currently understood are sufficient for this context remains an open question. But they're at least asking the right questions: verify everything, trust nothing by default, assume compromise, limit damage when it occurs.
The real test comes when organizations move beyond pilot projects and deploy these systems at scale, with real money and real consequences attached. We'll find out then whether the frameworks we're building now were adequate or whether we're learning these lessons the expensive way.
Bob Reynolds is Senior Technology Correspondent for Buzzrag
Watch the Original Video
Securing AI Agents with Zero Trust
IBM Technology
13m 33sAbout This Source
IBM Technology
IBM Technology, a YouTube channel launched in late 2025, has swiftly garnered a following of 1.5 million subscribers. The channel serves as an educational platform designed to demystify cutting-edge technological topics such as AI, quantum computing, and cybersecurity. Drawing on IBM's rich history of technological innovation, it aims to provide viewers with the knowledge and skills necessary to succeed in today's tech-driven world.
Read full source profileMore Like This
AI Agents That Work While You Sleep: The Next Shift
Cloud-based AI coding agents now run scheduled tasks overnight. A developer built a news monitoring system in one afternoon that never sleeps.
Can Cyber Training Keep Up with AI Threats?
Explore how immersive cyber training can tackle AI-driven threats and empower human defense.
AI Agents Are Getting God Mode—And That's a Problem
IBM's Grant Miller explains how AI agents with elevated permissions create security nightmares—and what actually works to prevent privilege escalation.
Anthropic Bet on Teaching AI Why, Not What. It's Working.
Anthropic's 80-page Claude Constitution reveals a fundamental shift in AI design—teaching principles instead of rules. The enterprise market is responding.