The most interesting shift in AI coding tools isn’t about making one agent smarter. It’s about running many agents in parallel.
The Tools Are Shipping Now
GitHub Copilot: Agent HQ + Specialized CLI Agents
GitHub Agent HQ launched on February 5, 2026. You can now assign the same GitHub issue to Copilot, Claude, and Codex simultaneously. Each agent works asynchronously, proposes a different solution, and you pick the approach that works best—all within your repository context. Available to Copilot Pro+ and Enterprise subscribers.
GitHub Copilot CLI now ships with four specialized agents that can parallelize work:
- Explore — Navigate and understand codebases at speed
- Plan — Design solutions and map dependencies
- Task — Execute builds, tests, and commands
- Code-review — Critique changes with high signal-to-noise ratio
These agents can run simultaneously on different parts of your repository. One explores dependencies while another plans the refactor. One writes tests while another reviews the implementation.
Copilot CLI Fleets takes this further. In /experimental mode, the /fleet command dispatches parallel subagents to implement your plan. The key innovation: a SQLite database per session models dependency-aware tasks and TODOs. Agents don’t just run in parallel—they understand which tasks depend on which, avoiding conflicts and respecting execution order where it matters.
Claude Code: Agent Teams
Claude Code’s experimental Agent Teams take a different approach. Rather than competing agents, Claude creates a coordinated team. The architecture has four components:
- Team lead — Your main Claude Code session that creates the team, spawns teammates, and coordinates work
- Teammates — Separate Claude Code instances that each work on assigned tasks, each with its own context window
- Shared task list — Work items with dependency tracking that teammates claim and complete. File locking prevents race conditions when multiple teammates try to claim the same task
- Mailbox — A messaging system where teammates communicate directly with each other, not just back to the lead
This last point is the key distinction from Claude’s simpler “subagents” feature. Subagents report results back to the caller. Agent team members message each other directly—sharing findings, challenging each other’s assumptions, and self-coordinating through the shared task list.
Enable it via the CLAUDE_CODE_EXPERIMENTAL_AGENT_TEAMS environment variable, then describe the team you want in natural language:
Create an agent team to review PR #142. Spawn three reviewers:
- One focused on security implications
- One checking performance impact
- One validating test coverage
The lead creates the team, spawns teammates, and synthesizes findings. You can interact with individual teammates directly (Shift+Up/Down in-process, or via split panes with tmux/iTerm2). Delegate mode (Shift+Tab) restricts the lead to coordination only—no coding—so it focuses purely on orchestration.
Best use cases from the docs: research and review with parallel investigation, new modules where teammates own separate pieces, debugging with competing hypotheses where agents challenge each other’s theories, and cross-layer coordination spanning frontend/backend/tests.
The distinction between Copilot’s approach and Claude’s is worth noting. Copilot’s Agent HQ treats agents as competing approaches—you compare and pick the winner. Claude’s Agent Teams treat agents as collaborating teammates—they coordinate, communicate, and build on each other’s work. Both models have merit depending on the task.
Local Swarms: Skip the Cloud Entirely
What if you want agent swarms without API costs, rate limits, or sending code to the cloud?
Developers are running Claude Code’s agent orchestration against local models. Here’s one developer’s setup running 4 agents locally on Qwen3-Coder:
I just woke up Claude Code Agent Swarm on local Qwen3 Coder Next. No cloud. No Internet. No quota anxiety. No ‘You’ve hit your limit, resets 10 pm’
By the numbers:
- ~100 tokens/sec generation
- 17,871 tokens/sec read top speed
- 256k context window
- Swarm tool calling just works™ out-of-the-box
My AI “data center” 1.25kg, 15x15x5.1 cm is now running autonomous coding agents faster than most SaaS APIs. Token cost is zero. Latency is sub-100ms. Vendor lock-in is deprecated.
The key insight: the orchestration patterns (Agent Teams, parallel execution, shared task lists) work the same whether the underlying model is cloud-hosted or local. Ollama lets you run models like Qwen3-Coder locally, and Claude Code can use them as the backend.
Cloud-based agent swarms (GitHub Agent HQ) offer seamless integration and access to frontier models. Local swarms offer speed, cost control, and privacy. These aren’t competing approaches—they’re complementary.
The Real Constraint: Orchestration
The limiting factor isn’t compute anymore. It’s orchestration intelligence.
When you run multiple agents in parallel, new problems emerge:
- How do you prevent conflicting edits to the same file?
- When do you run agents sequentially (one builds on another’s work) versus in parallel (independent explorations)?
- How do you merge insights from multiple agent approaches into one coherent solution?
- What’s the decision framework for picking which agent’s output to ship?
GitHub Agent HQ handles this by letting agents compete—you choose the winner. Claude’s Agent Teams handle it through coordination—the lead delegates, teammates communicate directly, and the shared task list tracks dependencies. Copilot CLI Fleets use a SQLite database for dependency-aware scheduling. Local setups require more manual orchestration, though the patterns are converging.
What This Means for Your Workflow
Pull requests reviewed by multiple AI perspectives before human eyes see them. One agent checks security, another checks performance, a third checks maintainability. All in parallel.
Faster exploration of solution spaces. Instead of iterating with one assistant through 5 approaches sequentially, spawn 5 agents exploring in parallel. Minutes later, you have 5 complete implementations to evaluate.
Specialization over generalization. Generic “do everything” assistants get supplemented by teams of specialists. One excels at test generation. Another at refactoring. Another at documentation. You orchestrate the specialists.
Getting Started
GitHub Copilot Pro+ or Enterprise:
- Try Agent HQ on your next complex issue
- Assign the same issue to multiple agents, compare their approaches
- Use Copilot CLI’s specialized agents for parallel workflows
- Try
/fleetin/experimentalmode to dispatch dependency-aware parallel subagents
Claude Code:
- Enable agent teams: set
CLAUDE_CODE_EXPERIMENTAL_AGENT_TEAMS=1in your environment orsettings.json - Describe the team you want in natural language—Claude spawns teammates based on your prompt
- Use delegate mode (Shift+Tab) to keep the lead focused on coordination
- Start with research or review tasks before attempting parallel implementation
We went from one AI assistant to an AI engineering team. The tools are shipping now. The question is how quickly orchestration patterns mature.
I’d love to hear your experience with multi-agent workflows. Connect with me on LinkedIn to continue the discussion.