Skip to content
Go back

Delegating to AI Agents Is First-Time Management Again (With One Brutal Difference)

Published:  at  10:00 AM

Here’s something I told my engineering teams recently: using AI agents is like learning delegation.

Imagine you just hired eight junior developers. Every thirty minutes, each one walks up to your desk and says: “Done. Can you review this?” That’s sixty-four review requests a day. No manager could survive that. You’d either burn out or stop delegating altogether.

Now replace “junior developers” with “AI agents.” Same problem. If your agents constantly need a human to route, validate, and decide what’s next, you haven’t delegated. You’ve built an interruption machine.

What surprised me is that most of the struggle teams face with agents has little to do with the technology. It has to do with delegation itself — a skill that managers have been learning (and failing at) for decades.


Every new manager has been here before

Every first-time manager goes through the same painful shift: from doing to getting outcomes through others. The management literature has documented the same barriers for fifty years. The need for control. The illusion that explaining takes longer than doing it yourself. The discomfort of seeing someone else produce something different from what you would have done.

Developers facing AI agents hit every single one of these barriers.

The developer who rewrites the agent’s output line by line is the exact equivalent of the manager who redoes their team’s work every evening. The developer who says “it’s faster to code it myself” is echoing every new manager who ever refused to delegate a task they could do in thirty minutes.

The good news: the progression is also well-known. You start by giving detailed instructions and reviewing everything. Then you provide richer context, better examples, and review less. Eventually you set high-level goals, put guardrails in place, and only intervene on exceptions. It maps almost directly to how teams mature in their use of agents.

And there’s one thing that makes delegation to agents structurally simpler than managing people: no human complexity. No career expectations. No office politics. No emotional management. That’s one less dimension to worry about.

But there’s another dimension that makes it brutally harder.


The bottleneck moved

With AI agents, production becomes nearly instant. An agent generates in minutes or hours what used to take hours or weeks.

The bottleneck shifts to validation. And validation is the one task that requires the most expertise. It demands the very judgment that took years to build — the ability to look at a piece of work and sense whether it will hold under real conditions. You can delegate production. You cannot easily delegate that judgment.

Worse: agent output often looks right. A human junior who’s unsure will hesitate, ask a question, flag a doubt. An agent delivers confident nonsense with the same formatting as its best work. The risk isn’t that agents produce bad code. It’s that overwhelmed humans validate it badly.

This is why the real skill is what I’d call delegation design: a clear definition of done, explicit constraints and guardrails, identified decision points where the agent should stop and ask, and a verification strategy — tests, rules, review targets. If the brief is vague, the agent will fill the gaps with confident noise. Just like a junior developer would — except faster and without raising their hand.


Two ways out of the bottleneck

If you can’t review everything, you need to change the equation. There are two paths, and they work best together.

Path one: raise autonomy. Give agents larger, better-scoped missions and invest heavily in the trust infrastructure. The stronger the safety net, the less the human needs to review. You intervene on exceptions only, like a manager who trusts their team but watches the dashboards.

Path two: add a supervisory layer. A “senior” agent pre-validates the work of “junior” agents — checking consistency, standards, coverage, patterns. It escalates to the human only what requires genuine judgment: ambiguity, tradeoffs, decisions.

The human’s role shifts from reviewer-of-every-output to architect-of-the-system-that-reviews. That’s a higher-order skill, not a lesser one.

One thing to keep in mind: unreviewed agent output has zero value. If it piles up waiting for a human who can’t keep up, your delegation model is broken — not your agents.

But let’s be explicit about the limits. High-stakes decisions still need human sign-off. Agent permissions must be tightly controlled. Domain correctness — especially in vertical B2B software where business rules are dense, subtle, or worse, tacit — needs explicit rules and tests, not hope. If you ignore these, you get one of two failure modes: blanket rejection (“too risky, let’s not use agents”) or blind trust (“looks good, ship it”).


Nobody trained your team for this

The expertise doesn’t vanish. It changes form. Right now, the best people to supervise agents are those who’ve spent years writing code themselves — they know what good looks like because they’ve built it. But that won’t last forever. When factories mechanized, the first foremen were former craftsmen. Within a generation, a new kind of expertise emerged — people who’d learned to operate, monitor, and optimize machines without ever having worked the metal by hand. The same shift will happen here. The question is whether we’re ready to build that new expertise deliberately, or whether we’ll just let people figure it out on their own.

But here’s the problem nobody is solving yet. When a company promotes someone to manager, there’s onboarding, training, mentoring. When an entire engineering team needs to become agent managers overnight — there’s nothing. No curriculum, no playbook, no support structure. And unlike a traditional promotion, it’s not one person making the transition. It’s everyone, at the same time.

Agentic AI won’t deliver on its promise until engineering teams treat delegation and verification as a first-class skill — not something you bolt onto existing processes.

So I’ll end with a question. Are you still reviewing everything your agents produce? Or have you started building the autonomy and verification layers that actually let this scale?



Next Post
Agentic AI Is Fundamentally Changing the Role of Tech Teams