The most interesting shift in AI coding tools isn’t about making one agent smarter. It’s about running many agents in parallel.

The Tools Are Shipping Now

GitHub Copilot: Agent HQ + Specialized CLI Agents

GitHub Agent HQ launched on February 5, 2026. You can now assign the same GitHub issue to Copilot, Claude, and Codex simultaneously. Each agent works asynchronously, proposes a different solution, and you pick the approach that works best—all within your repository context. Available to Copilot Pro+ and Enterprise subscribers.

GitHub Copilot CLI now ships with four specialized agents that can parallelize work:

  • Explore — Navigate and understand codebases at speed
  • Plan — Design solutions and map dependencies
  • Task — Execute builds, tests, and commands
  • Code-review — Critique changes with high signal-to-noise ratio

These agents can run simultaneously on different parts of your repository. One explores dependencies while another plans the refactor. One writes tests while another reviews the implementation.

Copilot CLI Fleets takes this further. In /experimental mode, the /fleet command dispatches parallel subagents to implement your plan. The key innovation: a SQLite database per session models dependency-aware tasks and TODOs. Agents don’t just run in parallel—they understand which tasks depend on which, avoiding conflicts and respecting execution order where it matters.

Claude Code: Agent Teams

Claude Code’s experimental Agent Teams take a different approach. Rather than competing agents, Claude creates a coordinated team. The architecture has four components:

  • Team lead — Your main Claude Code session that creates the team, spawns teammates, and coordinates work
  • Teammates — Separate Claude Code instances that each work on assigned tasks, each with its own context window
  • Shared task list — Work items with dependency tracking that teammates claim and complete. File locking prevents race conditions when multiple teammates try to claim the same task
  • Mailbox — A messaging system where teammates communicate directly with each other, not just back to the lead

This last point is the key distinction from Claude’s simpler “subagents” feature. Subagents report results back to the caller. Agent team members message each other directly—sharing findings, challenging each other’s assumptions, and self-coordinating through the shared task list.

Enable it via the CLAUDE_CODE_EXPERIMENTAL_AGENT_TEAMS environment variable, then describe the team you want in natural language:

Create an agent team to review PR #142. Spawn three reviewers:
- One focused on security implications
- One checking performance impact
- One validating test coverage

The lead creates the team, spawns teammates, and synthesizes findings. You can interact with individual teammates directly (Shift+Up/Down in-process, or via split panes with tmux/iTerm2). Delegate mode (Shift+Tab) restricts the lead to coordination only—no coding—so it focuses purely on orchestration.

Best use cases from the docs: research and review with parallel investigation, new modules where teammates own separate pieces, debugging with competing hypotheses where agents challenge each other’s theories, and cross-layer coordination spanning frontend/backend/tests.

The distinction between Copilot’s approach and Claude’s is worth noting. Copilot’s Agent HQ treats agents as competing approaches—you compare and pick the winner. Claude’s Agent Teams treat agents as collaborating teammates—they coordinate, communicate, and build on each other’s work. Both models have merit depending on the task.

Local Swarms: Skip the Cloud Entirely

What if you want agent swarms without API costs, rate limits, or sending code to the cloud?

Developers are running Claude Code’s agent orchestration against local models. Here’s one developer’s setup running 4 agents locally on Qwen3-Coder:

I just woke up Claude Code Agent Swarm on local Qwen3 Coder Next. No cloud. No Internet. No quota anxiety. No ‘You’ve hit your limit, resets 10 pm’

By the numbers:

  • ~100 tokens/sec generation
  • 17,871 tokens/sec read top speed
  • 256k context window
  • Swarm tool calling just works™ out-of-the-box

My AI “data center” 1.25kg, 15x15x5.1 cm is now running autonomous coding agents faster than most SaaS APIs. Token cost is zero. Latency is sub-100ms. Vendor lock-in is deprecated.

Mitko Vasilev

The key insight: the orchestration patterns (Agent Teams, parallel execution, shared task lists) work the same whether the underlying model is cloud-hosted or local. Ollama lets you run models like Qwen3-Coder locally, and Claude Code can use them as the backend.

Cloud-based agent swarms (GitHub Agent HQ) offer seamless integration and access to frontier models. Local swarms offer speed, cost control, and privacy. These aren’t competing approaches—they’re complementary.

The Real Constraint: Orchestration

The limiting factor isn’t compute anymore. It’s orchestration intelligence.

When you run multiple agents in parallel, new problems emerge:

  • How do you prevent conflicting edits to the same file?
  • When do you run agents sequentially (one builds on another’s work) versus in parallel (independent explorations)?
  • How do you merge insights from multiple agent approaches into one coherent solution?
  • What’s the decision framework for picking which agent’s output to ship?

GitHub Agent HQ handles this by letting agents compete—you choose the winner. Claude’s Agent Teams handle it through coordination—the lead delegates, teammates communicate directly, and the shared task list tracks dependencies. Copilot CLI Fleets use a SQLite database for dependency-aware scheduling. Local setups require more manual orchestration, though the patterns are converging.

What This Means for Your Workflow

Pull requests reviewed by multiple AI perspectives before human eyes see them. One agent checks security, another checks performance, a third checks maintainability. All in parallel.

Faster exploration of solution spaces. Instead of iterating with one assistant through 5 approaches sequentially, spawn 5 agents exploring in parallel. Minutes later, you have 5 complete implementations to evaluate.

Specialization over generalization. Generic “do everything” assistants get supplemented by teams of specialists. One excels at test generation. Another at refactoring. Another at documentation. You orchestrate the specialists.

Getting Started

GitHub Copilot Pro+ or Enterprise:

  • Try Agent HQ on your next complex issue
  • Assign the same issue to multiple agents, compare their approaches
  • Use Copilot CLI’s specialized agents for parallel workflows
  • Try /fleet in /experimental mode to dispatch dependency-aware parallel subagents

Claude Code:

  • Enable agent teams: set CLAUDE_CODE_EXPERIMENTAL_AGENT_TEAMS=1 in your environment or settings.json
  • Describe the team you want in natural language—Claude spawns teammates based on your prompt
  • Use delegate mode (Shift+Tab) to keep the lead focused on coordination
  • Start with research or review tasks before attempting parallel implementation

We went from one AI assistant to an AI engineering team. The tools are shipping now. The question is how quickly orchestration patterns mature.

I’d love to hear your experience with multi-agent workflows. Connect with me on LinkedIn to continue the discussion.

Updated: