12 min read

User ↔ Agent ↔ Swarm: The Winning Pattern

I used to think the goal was fully autonomous agents — AI that could work for hours without supervision. Then I actually started using AI agents seriously. The winning pattern isn't autonomy. It's orchestration.

Kyle Rasmussen
January 20, 2026
User ↔ Agent ↔ Swarm: The Winning Pattern
User ↔ Agent ↔ Swarm: The Winning Pattern

"Every single nine is the same amount of work."

That's Andrej Karpathy (@karpathy) — former OpenAI co-founder, Tesla's AI director — explaining to Dwarkesh Patel why the AI agent hype is running way ahead of reality. When your demo works 90% of the time, that's just the first nine. You need a second nine. A third nine. A fourth. A fifth.

At Tesla, after five years of work on self-driving, they got through "maybe three nines." There are still more to go.

Here's the uncomfortable math that nobody wants to talk about: 99% reliability sounds amazing. Until you do 100 things. Then you're at 37%.

That's not pessimism. That's compound probability. And it explains why the gap between AI demos and AI products feels so vast.

I used to think the goal was fully autonomous agents — AI that could work for hours without supervision, completing complex tasks while I slept. That was the dream, right? Set it and forget it. The machine does the work.

Then I actually started using AI agents seriously. And I learned something the demos don't show you: the winning pattern isn't autonomy — it's orchestration.


The Autonomy Fantasy

Everyone in AI right now is chasing the same thing: agents that work unsupervised.

You've seen the demos. You've probably seen the viral tweet — Cursor's team ran hundreds of concurrent agents for nearly a week, producing over a million lines of code and building a web browser from scratch. A browser! Autonomous AI! The future is here!

Except... read the fine print. They used a "planner-worker" hierarchy with dedicated coordination roles. They found that flat self-coordination failed — agents became "risk-averse" and "churned for long periods without progress." They needed periodic fresh starts "to combat drift and tunnel vision." And the code? Still "needs careful review."

That's not autonomy. That's orchestration with very long leashes.

Three different communities are tackling this problem right now, each with their own philosophy:

Ralph Wiggum (Geoff Huntley, @GeoffreyHuntley): A deceptively simple while loop — while :; do cat PROMPT.md | agent ; done — that built an entire programming language. The insight: context engineering and declarative specs can make "dumb things work surprisingly well."

Agent Flywheel (Jeffrey Emanuel, @doodlestein): Full VPS infrastructure for running 10+ agents 24/7. Tools for coordination (NTM, Agent Mail, BEADS) but the human still orchestrates. Investment: ~$500/month in subscriptions.

Gas Town (Steve Yegge, @steve_yegge): A full orchestration system with agent hierarchies — Mayor, Polecats, Refinery, Witness. His warning: "Only for Stage 7+ users who already juggle 5+ Claude Codes daily."

xAI Macrohard (Sulaiman Ghori, @sulaimanghori): The extreme end of the spectrum — "human emulators" that do anything a person does at a computer. xAI tested these internally as virtual employees, leading to surreal moments: staff tried to meet with colleagues who turned out to be AI. The ambition? Deploy millions of these emulators on idle Tesla vehicle computers. Notably: even xAI keeps humans discovering and debugging these systems. The emulators don't manage themselves.

The common thread? Every single one keeps humans in the loop. Ralph requires you to write the specs. Flywheel requires you to direct the coordination. Gas Town explicitly warns you'll need "a lot of elbow grease" to keep it running. And even xAI's most ambitious human emulators still need humans to discover their use cases and debug when things go sideways.


But let's walk through the math that explains why.

Let's say you have an exceptionally good AI agent — 99% reliable on each individual step. That's world-class. That's better than most human employees on routine tasks.

Now ask that agent to complete something that requires 100 steps. Maybe it's building a feature, or processing a complex workflow, or coordinating across multiple systems.

0.99 ^ 100 = 0.366

Thirty-seven percent success rate. On something that works 99% of the time per step.

This is what Karpathy calls the "march of nines" — and it's why there's a massive gap between AI that demos well and AI that actually works in production. The demo is easy. The product is hard.

Now consider that real-world tasks often involve hundreds or thousands of decision points. The math gets brutal fast.

This isn't a temporary limitation. This is a fundamental property of compound systems. Every additional step, every additional decision, every additional handoff multiplies the failure modes. And unlike human workers, AI agents can fail in spectacular, unpredictable ways — what Karpathy describes as "jagged intelligence." They're simultaneously smarter than expected on some things and shockingly dumb on others.

So what do the companies actually deploying AI agents do? They add human checkpoints. They build in review loops. They keep humans in the critical path.

The "autonomy" is marketing. The reality is orchestration.


The Conductor Insight

Here's where things get interesting.

Addy Osmani (@addyosmani), engineering leader at Google, wrote a piece in early 2026 that finally gave language to what I'd been experiencing. He describes the evolution of the developer role:

The role of the software engineer is evolving from implementer to manager, or in other words, from coder to conductor and ultimately orchestrator.

Think about what that means. Up to 90% of software engineers are now using some kind of AI for coding. But they're not being replaced — they're being elevated. From writing every line themselves to directing AI assistants. From individual contributors to conductors of an AI ensemble.

And here's the key insight: if a conductor works with one AI "musician," an orchestrator oversees an entire symphony of multiple AI agents working in parallel.

The conductor metaphor clicks something into place. When an orchestra gets better — more talented musicians, better instruments — does the conductor become less important?

No. They become more important.

A student orchestra needs a conductor to keep time. A world-class orchestra needs a conductor to bring out nuance, to make split-second interpretive decisions, to coordinate complexity that individual musicians couldn't manage alone.

The better the musicians, the more essential the conductor.

The same is true for AI agents. As they become more capable, they can do more — which means there's more to coordinate, more to direct, more potential for things to go sideways in subtle ways.

The winning pattern isn't removing the human. It's elevating the human to conductor.


My Swarm Journey

Let me make this personal.

Six months ago, I was using AI the way most people do: one conversation at a time. Now I regularly run 10-agent swarms that produce content, research, and code in parallel.

The shift happened in stages, and the revelation was where the human effort actually belongs.

The Planning Phase: Where the Real Work Happens

Here's what the demos don't show you: most of my "swarm time" isn't during execution. It's before.

I spend hours in user-agent pairing mode — one-on-one with a planning agent — building what I call a "robust plan." We work through known unknowns, validate assumptions, identify edge cases. The plan breaks work into granular tasks, each with clear acceptance criteria.

This isn't optional. A mediocre plan produces mediocre swarm output. A robust plan — one that anticipates failure modes and specifies exactly what "done" looks like — produces work you can actually use.

The planning agent and I go back and forth. "What if this assumption is wrong?" "What are the dependencies between these tasks?" "What does success look like for task 7?" By the time we're done, the swarm has a detailed map with 30+ specific tasks.

The Launch Phase: Currently Abrasive

I won't pretend launch is smooth. Right now, getting agents properly configured — right models, right context, right initial messages — requires attention. Agents need to "kick off" correctly, or they spend their first 10 minutes confused.

This will get easier with better tooling. But today, launch is hands-on. Think of it like a pit crew at a race — intense, focused, then they step back.

The Execution Phase: Coordinator-Led, Not Hand-Held

Here's the surprise: during execution, I'm mostly not involved.

The key is the Coordinator — a single agent whose job is to manage the swarm. The Coordinator constantly checks the task list, checks in with individual agents, and handles the orchestration I would otherwise do manually.

More importantly, the Coordinator seeds exploration. "Think more creatively about the tasks you just completed." "Are there adjacent problem spaces we didn't initially identify?" It nudges agents to go beyond the plan when appropriate.

When I do intervene, it's strategic: a blocker the Coordinator can't resolve, or a direction change based on early results. But I'm not hand-holding. The Coordinator earns its title.

The Review Phase: Structured Synthesis

This is where the hard problems live.

How do you efficiently review what a 10-agent swarm produced? That's 10 different workstreams, each with multiple outputs. Reading everything defeats the purpose of parallelization.

Here's my current approach: the Coordinator and I co-create a swarm review report. This isn't a list of outputs — it's a structured synthesis:

  • Plan vs. Reality: What was the original plan? Which tasks completed? Where did we drift, and was that drift productive or problematic?
  • BEADS Reconciliation: Every task in the system gets accounted for. Closed, carried forward, or discovered mid-swarm.
  • Reviewer Findings: If a Reviewer agent was part of the swarm, their critiques get spotlighted — what they flagged, what was addressed, what remains open.
  • Unexpected Discoveries: Sometimes agents find adjacent problems the plan didn't anticipate. These get surfaced explicitly.

The report compresses a 10-agent swarm's output into something I can review in 15 minutes. Without it, I'd be reading raw output for hours — which defeats the entire point of parallelization.

This is synthesis, not validation. Before anything goes to production, the standard gates still apply — unit tests for code, evals for AI outputs, fact-checking for content. The review tells me what to test; it doesn't replace the testing.


The Trust Layer

Here's the counterintuitive argument I want you to consider: human-in-the-loop isn't a bug — it's the feature.

The AI industry frames HITL as a temporary crutch. "Sure, you need human oversight now, but once the models get good enough..." The implication is that humans in the loop are a limitation to be engineered away.

I think that's exactly backwards.

Simon Willison (@simonw), one of the most thoughtful practitioners writing about LLMs, puts it bluntly: there's one key feature that remains unique to human staff: accountability.

A human can take responsibility for their actions and learn from their mistakes. Putting an AI agent on a performance improvement plan makes no sense at all!

This isn't new wisdom. IBM knew it 45 years ago. In 1979, a training slide stated: "A computer can never be held accountable. Therefore a computer must never make a management decision."

The trust problem isn't just about whether AI makes mistakes. It's about what happens when it makes mistakes. Who's responsible? Who learns? Who adjusts?

The numbers tell the story. The World Economic Forum and Capgemini's "AI Agents in Action" found that 82% of executives plan to adopt AI agents within three years. But BCG's "Widening AI Value Gap" research shows the reality for companies already using AI: only 5% are achieving value at scale. The majority — 60% — report minimal returns despite substantial investment.

That gap? That's the trust problem. Not trust in AI capabilities — trust in the system.

The companies in that 5% aren't the ones pursuing maximum autonomy. They're the ones who've figured out the orchestration layer — what BCG describes as reshaping work around humans and AI together, not replacing humans with AI.

Notice what's in that list: augmented human workers. Not replaced. Augmented.

Human-in-the-loop isn't the limitation holding AI back. It's the architecture that makes AI safe to deploy at scale.

But here's what the AI optimists miss: even when the technology is ready, the organizations won't be.

Put yourself in the shoes of a CEO. Your AI system just made a decision you didn't expect. Maybe it wasn't wrong — maybe it was actually clever. But you can't explain why it did what it did. You can't predict what it will do next. Are you comfortable with that?

Most leaders aren't. And they shouldn't be.

The path from "AI can do this" to "we trust AI to do this" isn't a technology problem — it's a change management problem. It requires:

  • Visibility: Leaders need to see how decisions are made, not just outcomes
  • Predictability: The system should behave consistently with stated principles
  • Reversibility: Mistakes need to be catchable before they cascade
  • Accountability: Someone human needs to own the result

This is why the orchestration skills matter so much. The companies that master user-agent pairing today — that build muscle memory for directing AI, for recognizing drift, for knowing when to intervene — those are the companies that will be ready when the technology catches up to the ambition.

The transition won't be "flip a switch to full autonomy." It will be a gradual expansion of the leash, earned through demonstrated reliability, measured in nines.


What This Means for You

If you're reading this and thinking "okay, but what do I actually do?" — here's the shift:

Stop trying to remove yourself from the process. Start trying to become a better conductor.

  1. Start with one agent, master the interaction. Don't jump straight to swarms. Learn to work alongside a single AI agent first. Understand its patterns. Learn to recognize when it's drifting off course. Build intuition for intervention.

  2. Watch the work happen. This is the mistake most people make — they send prompts and check results. The magic is in watching the agent work. You'll start recognizing patterns. You'll learn to intervene at the right moments. You'll develop the Matrix vision.

  3. Scale to parallel tasks. Once you can reliably direct one agent, try two. Then three. Feel what it's like to coordinate multiple workstreams. The cognitive shift is real — you'll know when you've made it.

  4. Keep yourself in the critical path. This isn't about micromanagement. It's about strategic checkpoints. Where are the decisions that require judgment? Where are the failure modes that could cascade? That's where you stay.

  5. Embrace the conductor role. The orchestra doesn't resent the conductor — they rely on them. Your agents don't need you out of the way. They need you directing traffic.


The Winning Pattern

So here's the pattern that actually works:

User ↔ Agent ↔ Swarm

You at the center. Agent(s) as your instruments. The swarm as your amplifier.

Clear separation of concerns. The agents handle execution — they're fast, tireless, capable of parallel work at scale. You handle coordination, judgment, and the decisions that require accountability.

The agent doesn't replace you. The swarm doesn't make you obsolete. They make you more powerful — but only if you step into the conductor role.

Pure agentic AI — agents that work for hours unsupervised, that complete complex tasks without human involvement — that's a fantasy for 2026. The math doesn't work. The accountability doesn't work. The trust doesn't work.

But human-orchestrated AI? Agents that amplify human judgment instead of replacing it? Swarms that multiply what a single person can accomplish?

That's not fantasy. That's what's working right now.

The question isn't whether AI agents will be able to work autonomously someday. Maybe they will. The march of nines continues, and every year brings new capabilities.

The question is: what do you do today?

My answer: become the conductor.

The Ralph Wiggum folks will tell you it starts with great specs. The Flywheel community will tell you it starts with the right infrastructure. Steve Yegge will tell you it starts with the right molecular workflow system.

They're all right. And they're all building variations on the same theme: humans orchestrating AI, not AI replacing humans.

The orchestra is ready. The instruments are tuned. The only thing missing is someone to direct the symphony.



Sources

  • Karpathy, Andrej (@karpathy). Interview with Dwarkesh Patel. "Every single nine is the same amount of work."
  • Cursor Research. "Scaling long-running autonomous coding." January 2026. Details on multi-week agent experiments.
  • Osmani, Addy (@addyosmani). "The Future of Agentic Coding." On the evolution from coder to conductor to orchestrator.
  • Willison, Simon (@simonw). "AI Agents and Accountability." On why human accountability remains irreplaceable.
  • BCG & World Economic Forum. "The FAST Framework for AI-Driven Transformation." Enterprise AI adoption statistics.
  • Huntley, Geoff (@GeoffreyHuntley). "The Ralph Wiggum Technique." Simple loops, powerful results.
  • Emanuel, Jeffrey (@doodlestein). "Agent Flywheel." Infrastructure for multi-agent development.
  • Yegge, Steve (@steve_yegge). "Welcome to Gas Town." Full orchestration systems for AI swarms.
  • Ghori, Sulaiman (@sulaimanghori). "WTF is happening at xAI." Relentless podcast, January 2026. On Macrohard human emulators and internal testing.

Next up: Where Is the Agent Hub?

Also published on