Swarm Command supports three scaling configurations, from SS-50 (fast) to SS-250 (full consensus swarm). This document helps you choose the right scale quickly, then explains the details.
Need a fast second opinion on 1β2 files?
β SS-50
Need a strong answer for a subsystem, feature, or review?
β SS-100
Need repo-wide coverage, maximum consensus, or high-stakes analysis?
β SS-250
Is the task bounded to 1β2 files or one very narrow question?
ββ Yes β SS-50
ββ No
β
ββ Does it span a feature, module, or several files with real implementation risk?
β ββ Yes β SS-100
β ββ No
β
ββ Does it touch many modules, policy/compliance, or require maximum coverage?
ββ Yes β SS-250
ββ If unsure β start at SS-100 and scale up only when coverage matters more than speed
| Scale | Total Agents | Commanders | Squad Leads | Workers | Reviewers | Best For | Wall-Clock |
|---|---|---|---|---|---|---|---|
| SS-50 | ~36-52 | 2-3 | β | 30-45 | 3 | Fast bounded tasks | ~30s |
| SS-100 | ~89 | 5 | β | 75 | 8 | Default for real software work | ~45s |
| SS-250 | ~316 | 5 | 50 | 250 | 10 | Repo-wide or maximum-confidence work | ~65β90s |
Agent counts include all deployed agents across all layers: Nexus + Commanders + Squad Leads + Workers + Reviewers.
Default: SS-100. Use swarm command ss-250 for full deployment or swarm command ss-50 for quick work.
Best for: single-file refactors, focused investigations, quick code analysis, or doc updates for one component.
L0: 1 Nexus (claude-opus-4.6)
L1: 2-3 Commanders (commander pool β 9 models)
L2: 30-45 Workers (worker pool β 6 models) β 15 per commander, spawned directly
3 Reviewers (cross-family pairs, spawned by Nexus)
ββββββββββββββββββββββββββ
Total: ~36-52 agents
Cost: $1.50 β $3.50
Time: ~30s wall-clock
| Parameter | Value |
|---|---|
| Commanders | 2-3 |
| Domains covered | 2β3 of 5 (auto-selected by task type) |
| Squad Leads per Commander | β |
| Workers per Commander | 15 |
| Reviewers | 3 |
| Shadow scoring | Score computed, no hardening loop |
| Cost ceiling | $5.00 |
| Timeout cascade | 60/40/30/20s |
- β βTell me what this stack trace most likely meansβ
- β βReview this one file for perf or security issuesβ
- β βExplain this subsystem quickly so I can get unstuckβ
- β βUpdate docs for this single componentβ
- β Repo-wide or multi-module changes
Best for: multi-file features, module-level refactors, comprehensive reviews, and most day-to-day engineering tasks.
L0: 1 Nexus (claude-opus-4.6)
L1: 5 Commanders (commander pool β 9 models)
L2: 75 Workers (worker pool β 6 models) β 15 per commander, spawned directly
8 Reviewers (cross-family pairs, spawned by Nexus)
Shadow Scoring (Nexus-internal, sealed criteria)
ββββββββββββββββββββββββββ
Total: ~89 agents
Cost: $3.50 β $8.00
Time: ~45s wall-clock
| Parameter | Value |
|---|---|
| Commanders | 5 |
| Domains covered | All 5 |
| Squad Leads per Commander | β |
| Workers per Commander | 15 |
| Reviewers | 8 reviewers (drawn from 7 cross-family review pairs) |
| Shadow scoring | 8 sealed criteria, hardening at >15% |
| Cost ceiling | $10.00 |
| Timeout cascade | 75/50/35/25s |
- β Multi-file feature implementation
- β Module-level refactoring with tests and docs
- β Architecture review + implementation plan + rollout notes
- β API design, contract review, or comprehensive PR analysis
- β Tiny tasks where latency matters more than coverage
Best for: repo-wide refactors, full-feature implementation, comprehensive audits, and documentation or compliance overhauls where missing a category is expensive.
L0: 1 Nexus (claude-opus-4.6)
L1: 5 Commanders (commander pool β 9 models)
L2: 50 Squad Leads (claude-haiku-4.5 | gpt-5.4-mini) β 10 per commander
L3: 250 Workers (worker pool β 6 models) β 5 per squad lead
L4: 10 Reviewers (7 cross-family pairs, cycled to fill 10 slots)
Shadow Scoring (Nexus-internal, sealed criteria)
ββββββββββββββββββββββββββ
Total: ~316 agents
Cost: $8.00 β $16.22
Time: ~65β90s wall-clock
| Parameter | Value |
|---|---|
| Commanders | 5 |
| Domains covered | All 5 |
| Squad Leads per Commander | 10 |
| Workers per Squad Lead | 5 |
| Reviewers | 10 reviewers forming 7 cross-family pairs |
| Shadow scoring | 10 sealed criteria (SS-50: 6, SS-100: 8), hardening at >15% (SS-50: disabled) |
| Cost ceiling | $20.00 |
| Timeout cascade | 90/60/40/30s |
- β Repo-wide refactoring
- β Full feature implementation across many modules
- β Comprehensive security or compliance audit
- β Complete documentation overhaul
- β Multi-service architecture analysis
- β Simple tasks where the swarm would be mostly overhead
| Layer | Agents | Model | Tokens In (avg) | Tokens Out (avg) | Cost |
|---|---|---|---|---|---|
| Nexus (L0) | 1 | claude-opus-4.6 | 50K | 8K | $1.35 |
| Commanders (L1) | 5 | commander pool | 30K Γ 5 | 4K Γ 5 | $0.75 |
| Squad Leads (L2) | 50 | haiku / gpt-5.4-mini | 8K Γ 50 | 2K Γ 50 | $0.72 |
| Workers (L3) | 250 | worker pool | 2K Γ 250 | 0.5K Γ 250 | $0.90 |
| Reviewers (L4) | 10 | 7 cross-family pairs | 10K Γ 10 | 2K Γ 10 | $0.60 |
| Total | 316 | $4.32 (optimistic) |
Wall-clock time grows slower than agent count because the expensive work runs in parallel:
Agents Wall-Clock Ratio vs SS-50
50 ~30s 1.0Γ
100 ~42s 1.4Γ
250 ~65s 2.2Γ
These are design targets, not measured benchmarks. The serial bottlenecks are limited to:
- Nexus decomposition: ~2s
- Canary verification: ~3s
- Final synthesis: ~10s
Everything else overlaps via hierarchical fan-out and pipeline overlap.
| # | Strategy | Impact | Description |
|---|---|---|---|
| 1 | Use explore / task workers |
60% cheaper | Worker types are significantly cheaper than general-purpose |
| 2 | Haiku / Mini at L3 | 10Γ cheaper | Cheapest models handle the most atomic work |
| 3 | Micro-brief compression | ~15% savings | Smaller inputs reduce per-agent cost at scale |
| 4 | Wave deployment | ~20% savings on failure | Canary β Probe (max 3) β Remainder β with health gates between waves. Catches rate limits and bad tasks before full deployment. |
| 5 | Canary verification | ~5% savings on failure | One cheap canary prevents many expensive failures |
| 6 | Timeout cascade | Cost protection | Stop slow work before it burns budget |
| 7 | Cost ceiling | Absolute protection | $20 hard cap prevents runaway bills |
| Config | All-Claude | Mixed Claude+GPT | All-GPT | Budget (explore-heavy) |
|---|---|---|---|---|
| SS-50 | $3.50 | $2.80 | $2.50 | $1.50 |
| SS-100 | $8.00 | $6.50 | $5.50 | $3.50 |
| SS-250 | $16.22 | $12.50 | $10.00 | $8.00 |
Budget configuration uses 60% explore + 30% task + 10% general-purpose.
| Task Complexity | Files Touched | Domains Needed | Recommended Scale |
|---|---|---|---|
| Simple refactor | 1β2 | 1β2 | SS-50 |
| Module feature | 3β10 | 2β3 | SS-100 |
| Cross-module feature | 10β50 | 3β5 | SS-250 |
| Repo-wide migration | 50+ | 5 | SS-250 |
- Start at SS-100 when you want strong coverage but don't yet know whether you need maximum depth.
- Drop to SS-50 when latency is the main constraint.
- Jump straight to SS-250 when the cost of a missed class of issue is higher than the cost of waiting another 20β40 seconds.