8 min read
Barak Brudo

Your Deployment Pipeline Wasn’t Built for AI Agents – And It’s About to Show

Your pipeline was designed for human-authored code. AI agents don’t wait for humans. Here’s what that means for your infrastructure right now.

Deployment Pipeline

Deployment Pipeline

Imagine this scenario: a Kubernetes cluster shows signs of problems. No engineer opens a ticket. No engineer pages an on-call person. Instead, an automated triage agent is dispatched to investigate and respond to the problem. Not through human decision-making processes, but through real-time operational data.

This is not a future scenario. kagent, a CNCF sandbox project, is a working example of exactly this. Agents built on kagent can investigate cluster issues, diagnose problems, and take corrective action autonomously, reasoning over live logs, metrics, and cluster state, without waiting for a human to initiate the process.

That’s a problem that breaks more assumptions in your deployment pipeline than most engineering teams have had time to consider.

The conversation is happening at the wrong layer

Every infrastructure provider out there today is touting “AI Ready” in one way or another. Most of that discussion is happening on the model and tooling level, which models to utilize, which frameworks to construct on, how to orchestrate LLM calls, etc. That’s a necessary discussion, and it’s a good one, but it’s not the one that’s actually going to hurt us the most.

The one that isn’t happening, not loudly anyway, is what happens when an autonomous agent has to be deployed into an environment that’s constructed on a fundamentally different set of assumptions.

Gartner estimates that 40% of enterprise applications will be integrated with task-specific AI agents by the end of 2026, up from less than 5% at the end of 2025. That’s a very high growth curve, and it’s happening very, very fast. What it doesn’t address, however, is the infrastructure reckoning that’s happening as a result of it.

Your Kubernetes environment, deployment pipeline, secrets management, etc., were all designed with one assumption in mind: that somewhere in the pipeline, there’s a human being present, using judgment, using context, and absorbing the consequences of delay.

Agents don’t work that way.

What agents actually demand from your infrastructure

The mismatch is not abstract; it manifests itself in four ways:

Bursty, non-deterministic compute. Human-authored code is mostly deterministic. You can capacity plan for it. The agents are reactive; they start in response to events, consume compute in bursts, and may be running simultaneously on multiple clusters in ways that have little to do with your normal traffic patterns. The infrastructure you’ve designed for steady-state traffic will be under-provisioned for the bursts or over-provisioned all the time. Neither of these is a good plan when the agent is running.

Long-running stateful tasks. Most deployment pipelines are geared towards short-lived, stateless workloads. The agents have to maintain state over long periods of time, keep context, and interact with other systems in an iterative way over time. Workload schedulers that are geared towards transient work will interrupt agents in the middle of execution. Stateless assumptions are now bugs, not merely performance problems.

Permissions that span multiple systems at once. A kagent agent does not query a single system of record. It reasons over data from across your production environment, logs, metrics, and cluster state in real time. It needs access to all of this in real time. This is where your existing permissions framework was never designed to handle something like this: a non-human principal that needs broad access to multiple production systems at once without waiting for human approval at every step of the way.

Zero tolerance for deployment bottlenecks. When you deploy human-authored code and the deployment pipeline stalls, someone notices and fixes it. When an agent is running and waiting on a deployment to occur, the agent itself is now the bottleneck. A 20-minute deployment wait is not merely a process problem; it’s now a live failure with consequences that the agent cannot work around.

The governance gap that nobody’s budget covers

Deploying agents into production is dangerous enough as it is without adding infrastructure friction on top of that. This is where the infrastructure problem meets the security and compliance problem, and where the gap is widest. 

According to Cisco’s State of AI Security 2026, most organizations planned to deploy agentic AI into business functions, but only 29% reported being prepared to secure those deployments. That gap exposes risk across model interfaces, tool integrations, and supply chains simultaneously. Cisco’s AI Readiness Index puts a sharper number on it: only 24% of organizations have controls in place to govern agent actions with proper guardrails and live monitoring.

Those figures represent the number of organizations that are deploying, or are about to deploy, agents into infrastructure that is meant to be used for something else.

The problem of governance is different from the overall security discussion around AI. The problem of AI security is not “can I jailbreak the model?” The problem is “what if an agent that I’ve given access to make decisions somehow decides to do something my RBAC model wasn’t designed to prevent, uses a secret that my compliance team didn’t expect to be renewed within a certain timeframe, or initiates a workflow that my auditing wasn’t designed to capture?”

These are not edge cases. They are the default state of what it means to deploy autonomous systems into infrastructure whose model assumes a human would be present.

When the pipeline bottleneck becomes an agent bottleneck

There is a related problem, which is a compounder of everything above. That problem is tribal knowledge.

Any decent team of engineers will be intimately familiar with how their deployment pipeline actually works, the tricks they’ve learned along the way, the conditions under which it will appear to hang, the cases they’ve implemented manually because it would have taken longer to do it “properly.” Human engineers will use that knowledge to work around the pipeline problems that would stop an automated system cold.

The agents won’t have access to any of this knowledge. They will be working with what the infrastructure actually provides, rather than what an experienced engineer would know to work around. Every piece of tribal knowledge that hasn’t been implemented into your platform will be a source of failures that an agent will inevitably report, at the worst possible time, because agents don’t get to choose the timing.

A different Gartner analysis found that over 40% of agentic AI projects will be canceled by the end of 2027, with inadequate risk controls cited as a primary reason. That’s not a model problem. That’s an infrastructure and governance problem, one that surfaces when organizations move from proof of concept to production and discover that their environment wasn’t built for what they’re asking it to run.

The question we’re supposed to be asking

Nobody is building the wrong models. The models are getting better really fast. The tooling is legitimate.

The other question that’s getting less attention is: if an AI agent has to be deployed to production, if it has to have compute that scales to its demand, if it has to have state that persists across the session, if it has to have access to systems that it doesn’t own with appropriate access controls, if it has to have a deployment mechanism that doesn’t make it wait, what does your platform give it?

Not in theory. Not in a demo environment with relaxed guard rails. In production. Under load. With compliance requirements and consequences of failure.

How Control Plane is ahead of the curve 

Two features in particular speak directly to the challenges above.

Capacity AI™ addresses the bursty, non-deterministic compute problem head-on. Rather than forcing you to choose between chronic over-provisioning and being caught under-resourced when an agent fires, Capacity AI continuously analyzes real workload behavior and adjusts CPU and memory allocation in real time. Agents get the resources they need when they need them, and your infrastructure stops paying for headroom that sits idle the rest of the time.

Cloud Wormhole® addresses the multi-system access issue. The agents don’t just require access to one system; they require reasoning across logs, metrics, databases, and internal APIs, many of which reside inside VPCs or private networks that were intentionally kept off the public internet. Cloud Wormhole establishes secure, persistent tunnels to these private endpoints without forcing you to expose them or change your network topology. The agent gets the access it requires. Your security model remains unchanged.

These capabilities eliminate two of the four issues that make deploying agents into traditional environments an exercise in peril. The technology agents require to function in production is available. The question is whether your platform is ready to welcome agents with the infrastructure they need, or whether a production incident answers that question for you.