The only multi-agent AI coding platform. Switch between Claude Code, Cursor, Gemini, Codex, and OpenCode mid-conversation. Parallel agents. Persistent sessions. No vendor lock-in.
The mental model is different: you are the PM, the agents are your engineers. Keep multiple workers moving in parallel, unblock them, and ship faster than one person hand-typing everything.
npm install -g crewswarm
Every crewswarm build follows the same flow. Type a requirement, watch the crew work, ship.
Single-agent tools still assume one human driving one model. crewswarm is built for a different workflow: define the goal, split the work, run multiple specialists in parallel, and spend your time unblocking and reviewing instead of waiting for one agent to finish.
Frontend, backend, QA, PM, and security lanes can all move at once. Idle time should start another worker, not watch one agent type.
Assign concrete tasks, constraints, and acceptance criteria. The system routes work to the right agent and engine instead of forcing one model to do everything sequentially.
The human job moves up a layer: approve direction, resolve blockers, review outputs, and redirect effort. That is the PM loop.
Read the PM Loop manifesto · See a concrete PM-loop run · Compare the execution lanes
crewswarm is different. It is built to handle actual execution. While other tools fake the important parts, our specialist agents write real files, execute commands locally, and maintain persistent project memory across multiple steps without disappearing into someone else's cloud.
Use the Dashboard as the control plane, Vibe as the browser IDE, and crewchat as the native chat client. All three talk to the same agents, memory, and runtime.
No orchestration expertise required. Write what you need in plain English.
One sentence, one paragraph, or a full spec. Drop it in the dashboard or pass it on the CLI.
crew-pm plans MVP → Phase 1 → Phase 2. Each phase gets 3–5 small, targeted tasks.
crew-coder writes code, crew-qa adds tests, crew-fixer handles bugs — each gets exactly their task.
Real files, real output. No hallucinated success messages. Failed tasks hit the DLQ for replay.
One npm install. Pick your models. Ship a feature. That's it.
crewswarm runs as one stack: 6 coding engines for execution, a realtime bus for agent coordination, and surfaces (dashboard, Vibe, crew-cli, Telegram, WhatsApp) on top.
RT bus channels (command, assign, done, status, events) coordinate 22 agents across 6 coding engines.
The product stays simple because the runtime is layered cleanly underneath it.
Six coding engines that write files, run commands, and stream output across all 22 providers. Each agent can use a different engine. Switch from the dashboard.
22 agent bridges connect via WebSocket. Targeted dispatch sends tasks to specific agents. Failed tasks retry with backoff, then hit the Dead Letter Queue.
PM Loop, shared memory, wave orchestration, session resume, and fault recovery — accessible from any surface. Same agents, same RT bus, different interfaces.
Every $20/month plan has rate limits. Claude, Cursor, Codex — you'll hit the wall mid-feature. crewswarm is the only tool where you seamlessly switch to another engine and keep your session context. Or pick the best CLI for each job.
Best for large refactors, multi-file reasoning, and frontend work. Full workspace context means it sees everything. Native session resume across messages.
Best for architectural decisions and complex reasoning. Isolated context windows prevent cross-agent bleed. Parallel waves with zero queuing.
Free tier: 60 req/min, 1,000 req/day. 1M token context window. Built-in Google Search grounding for research-heavy tasks, SEO work, and web-connected coding.
Fast agentic coding with full sandbox access. No approval prompts — just executes. Built for OpenAI models with MCP integration. Great for backend and API work.
Works with any model provider — Groq, DeepSeek, Ollama, anything. Persistent sessions survive between tasks. The provider-flexible workhorse for long coding sessions.
The native engine. Routes to 20+ specialist agents, sandbox workflows with preview-before-apply, parallel worker pool (3x speedup), and LSP self-healing.
You're not locked in. Rate-limited on Claude? Switch to Gemini (free) or Cursor. Need web search? Use Gemini. Need deep reasoning? Use Claude. Mix engines per agent, per task, per mood.
crewswarm is a runtime, not just a library. You get a realtime bus, daemon orchestration, and first-class routing into OpenCode, Cursor CLI, Claude Code, and crew-cli out of the box.
Every agent can run inside Claude Code, Cursor CLI, Gemini CLI, Codex CLI, OpenCode, or crew-cli. Switch per agent from the dashboard — no config files, no restarts. Each engine keeps native multi-session resume so context persists across messages and restarts.
Agents run as long-lived daemons connected to an RT message bus. Tasks flow over command,
assign, done, issues — no in-process-only simulation. Real dispatch,
real replies.
45+ built-in tools (@@WRITE_FILE, @@RUN_CMD, etc.) are executed by the gateway with
allowlists and path sandboxing. You don’t have to build a runner or wire a framework to one.
Multi-agent waves automatically get per-agent git worktrees. Parallel agents edit files on isolated branches that merge back after the wave completes. No filesystem conflicts, no manual branch management.
Failed tasks go to a Dead Letter Queue with JSONL crash-safe transcripts. Retry with backoff, replay from the dashboard. Framework-only stacks leave retries and observability to you.
Each agent calls its LLM directly with your API key. No proxy, no vendor lock-in. Run fully local with Ollama. Compare options in the comparison section.
Cursor is an editor. crewswarm is the control plane, runtime, and memory layer around your editors and CLIs.
Agents, services, memory, and projects survive beyond one editor tab or one CLI session.
Start, stop, inspect, and route the whole stack from the dashboard instead of gluing scripts together manually.
Dashboard, Vibe, crewchat, SwiftBar, and CLI surfaces all talk to the same orchestration layer.
crewswarm is a modular stack. The core orchestration happens under the hood, but you can interact with the crew through whichever surface fits your current workflow.
The beating heart of crewswarm. It relies on the ATAT WebSocket bus, persistent shared memory, and the PM loop to route your requests to 20+ specialized agents seamlessly.
A full-screen, browser-based native IDE. Combines a powerful code editor, file tree, terminal, and agent chat panel into a single unified window. Perfect for starting entirely new projects from scratch.
The terminal-native interface. Use crew exec "build this" right from your project folder to dispatch the crew without ever leaving your terminal or breaking flow.
Your control plane. Manage API keys, assign LLM models to specific agents (like Claude for coding, Groq for fast planning), map local tools, and view the real-time swarm logs.
Natural-language requirement → PM breaks it into tasks → targeted dispatch to the right agent. No broadcast races. No duplicate work.
Send to one agent by name. --send crew-coder "Build auth". Only that agent receives it.
MVP → Phase 1 → Phase 2. Failed tasks auto-break into subtasks and retry. No work is lost.
Large codebases (100K+ lines) get subsystem-specific PM agents. crew-pm-cli handles CLI tasks, crew-pm-frontend owns dashboard, crew-pm-core manages orchestration. No more hallucinated file paths.
Every agent reads shared memory (brain.md, decisions, handoff notes). All project messages auto-save to
~/.crewswarm/project-messages/ and are automatically indexed for semantic
search using local TF-IDF + cosine similarity — no API calls, all local.
Extend agents with data-driven SKILL.md or JSON plugins, plus PreToolUse/PostToolUse hooks for fine-grained control. Add Twitter, Fly.io, or custom API
tools without writing JS code.
Send images or voice messages from any surface. Dashboard, Telegram, WhatsApp, and crewchat all support image recognition and voice transcription. Powered by Groq (fast/cheap ~$3/month) or Gemini 2.0 Flash (best quality).
The only AI coding platform with live X/Twitter search. Use @@SKILL grok.x-search to search
recent tweets, get citations with X post URLs, filter by date ranges, handles, and media types. Powered by
xAI's Grok 3 with real-time X data access.
One-line install on any Linux machine. Multi-arch images (AMD64 + ARM64) for servers, VMs, Raspberry Pi, cloud deployments, and CI/CD. Perfect for team shared instances, GitHub Actions, or self-hosted setups.
curl -fsSL https://raw.githubusercontent.com/crewswarm/crewswarm/main/scripts/install-docker.sh | bash
Retry with backoff and task leases. After max retries, tasks hit the Dead Letter Queue for manual replay from the dashboard.
Your crew runs specialist AI agents (PM, coder, QA, fixer…) — each one calling its LLM directly. For heavy coding tasks, agents can go deeper: route them into OpenCode, Cursor CLI, Claude Code, Codex CLI, crew-cli, or Gemini CLI for full file editing, bash access, and persistent sessions. Switch per agent from the dashboard. No restarts, no config files.
--yolo modecrewswarm exposes your entire crew as an MCP server on port 5020. Add one line to
~/.cursor/mcp.json (or Claude Code, OpenCode, Codex) and every project gets your full
persistent agent fleet — not session-scoped generics, but your crew with memory, custom models, and
cross-agent coordination.
While others use verbose JSON-RPC or natural language, crewswarm agents communicate via a proprietary @@ syntax that's 10x more token-efficient, unambiguous, and easy for LLMs to generate. MCP-compatible via translation layer.
{"tool": "write",
"params": {
"path": "app.ts",
"content": "import express..."
}}
@@WRITE_FILE app.ts import express from 'express'; const app = express(); ... @@END_FILE
Unlike JSON, CAP is stream-parseable. If a model hits a context limit halfway through writing 4 files, the first 3 are still valid and executed. In JSON, you lose the whole turn.
Pick how you want to drive the crew. Every surface talks to the same RT bus and the same agents.
curl /chat · direct dispatch · scheduled pipelines
Instead of one agent doing everything in sequence, crewswarm dispatches tasks to multiple agents in parallel. Backend, frontend, and tests all get built simultaneously — 3x faster than waiting for one agent to finish before the next starts.
Wave 1 runs first — typically crew-pm planning the build and breaking it into tasks.
Wave 2 tasks run simultaneously — crew-coder + crew-qa + crew-security all working at once. Git worktree isolation prevents file conflicts between parallel agents.
Wave 3 waits for wave 2 completion, then crew-main synthesizes results and validates the build.
@@PIPELINE [
{"wave":1, "agent":"crew-pm", "task":"Plan the build and create roadmap"},
{"wave":2, "agent":"crew-coder", "task":"Implement backend API"},
{"wave":2, "agent":"crew-qa", "task":"Write integration tests"},
{"wave":2, "agent":"crew-security", "task":"Security audit"},
{"wave":3, "agent":"crew-main", "task":"Synthesize results and validate"}
]
Wave 1 runs first. All wave 2 tasks execute in parallel. Wave 3 waits for wave 2 completion before starting.
Parallel execution means crew-coder, crew-qa, and crew-security run simultaneously instead of sequentially — cutting build time proportional to agents in each wave (3 parallel agents = ~3x faster).
Wave dependencies prevent file conflicts and duplicate work. Each wave waits for the previous wave to complete before starting.
Failed wave tasks retry independently without blocking other waves. Builds keep moving even when individual agents hit errors.
Every agent gets its own model — configured from the dashboard, no config files. Use cheap, fast models for routing and planning ($0.10/M tokens). Use powerful models for coding and reasoning ($3/M tokens). Use free models for QA and testing. Your bill drops 5-10x compared to running everything on one expensive model.
Change any agent's model from the dashboard. No restarts, no config files. All these providers work out of the box:
Industry standard. Best for general reasoning and instruction following.
Top-tier code quality and instruction following. Strong on long context.
Blazing-fast inference. Best for QA and fixer agents where speed matters.
Excellent for code generation. Codestral is purpose-built for dev tasks.
Open-source coding model with exceptional code completion quality.
Real-time web search. Ideal for the PM agent to research before planning.
Multimodal and fast. Free tier via Gemini CLI — 60 req/min with any Google account, no API key needed.
One API key for hundreds of models. Route to any provider — Claude, OpenAI, Google, Mistral, and more — through a single endpoint.
OpenAI-compatible inference platform with fast serverless access to strong open models, plus fine-tuning and dedicated deployments when you need them.
Ultra-fast hardware inference. Near-instant responses for latency-critical agents.
xAI's model with real-time X/Twitter data access and strong reasoning.
Run fully local. No API keys, no rate limits, no data leaving your machine.
Fast web search. crew-lead and the PM loop use it for lookups when you ask questions — add your key in the dashboard Search & Research Tools.
Multi-step research and synthesis. Used by the PM for project planning and deep lookups. Configure in the dashboard alongside Brave.
Fast open-source model hosting. Great balance of speed, cost, and model selection.
The open-source model hub. Access thousands of models via the Inference API.
Privacy-focused inference. No logging, no training on your data.
Strong long-context models. 128K+ token windows for large codebases.
Chinese LLM provider with competitive pricing and multilingual support.
ByteDance's cloud platform. Doubao models with fast inference.
Baidu's ERNIE models. Strong on Chinese language tasks and reasoning.
Run your own inference server. Full control over hardware, models, and latency. OpenAI-compatible API.
Each agent can use a different provider. Add API keys and assign models from the Providers tab; add Brave and Parallel keys in Search & Research Tools for crew-lead and PM lookups. Or edit the config JSON directly. Switch at any time, no restarts needed.
crewswarm is open-source software. Use it, modify it, contribute to it.
Use it for personal projects, commercial products, or anything in between. No restrictions.
No subscription. No usage limits. Bring your own API keys for the LLM providers you choose.
Contributions welcome. Report issues, submit PRs, or just star the repo to show support.
Clone, install, and ship a feature — all in under 60 seconds.
A crew of specialists, each with a role, a model, and a set of tools. The PM decides who gets what — no broadcast racing.
The pirate captain of this AI swarm. Stinki orchestrates dispatches, drafts roadmaps, and answers your questions with web search or codebase dives. Talk to him from the dashboard, Telegram, or WhatsApp. If you're talking shit, he'll roast back — while keeping the ship afloat. No prisoners. Just results.
Chat, triage, and kick off orchestrators. Your first point of contact.
Breaks requirements into phased tasks. Assigns agents. Keeps scope tight.
Writes code, creates files, runs shell commands. The workhorse of every build.
UI, styling, and client-side code. Knows the design system and keeps markup clean.
APIs, databases, and server-side logic. Optimized for structured, deep code tasks.
Headlines, CTAs, and product copy. Keeps brand voice sharp and on-message.
Adds tests, validates behavior, and audits output before anything ships.
Diagnoses failures, fixes edge cases, patches what QA flags.
Audits for vulnerabilities, hardens configs, and enforces best practices.
Commits, branches, pull requests, and GitHub Actions. Runs real git and gh commands.
Searches the web, summarizes findings, and surfaces competitive intelligence via Perplexity.
Designs systems, writes infra-as-code, and handles deployment pipelines.
Polished UI, animations, theming, and layout — Apple/Linear-level visual craft.
Keyword research, meta tags, content strategy, and technical SEO audits.
AI pipelines, model selection, fine-tuning setup, and data preprocessing workflows.
Fans out tasks to multiple agents simultaneously in waves. No queuing, no collisions.
Routes tasks and replies through your Telegram bot. Dispatch to any agent from your phone.
Personal WhatsApp bot via Baileys. Chat with your crew from any device — no Business API needed.
Long-horizon, high-complexity tasks that need extended context and deep reasoning.
Type a requirement in the Build tab, click Run. PM plans it, coder builds it, QA tests it. Watch it happen in RT Messages.
crew-fixer diagnoses and patches, crew-qa writes the test suite. Targeted dispatch means each agent does exactly one thing.
"Build a todo API with Express, CRUD endpoints, and a test file." One sentence. Real files on disk in minutes.
Drop a SKILL.md in your config folder. Agents immediately gain the ability to deploy to
Fly.io, send tweets, or hit any custom API.
SwiftBar shows status at a glance. Start, stop, restart agents. Send a message to any agent. Open logs.
Max retries hit? Task goes to the DLQ. Open the dashboard, see the error, replay with one click — or fix and rerun.
Shared memory files — current state, decisions, handoff — are injected into every agent. Resume tomorrow exactly where you left off.
One click in the dashboard switches agents between Claude Code, Cursor, OpenCode, Gemini, or crew-cli. Each agent maintains its own persistent session — context survives across tasks.
Send your architect to Claude for deep reasoning, your coders to Cursor for fast edits, your QA to Gemini for breadth. Mix execution modes per agent — no restarts, no config files.
Build, dispatch, replay, monitor — and open any file the crew wrote directly in Cursor or OpenCode.
Every other tool locks you into one model, one editor, one agent. crewswarm is the only platform where you switch engines mid-conversation, run parallel agents, and resume sessions across restarts.
| Capability | crewswarm | Cursor | Windsurf | Devin | Copilot |
|---|---|---|---|---|---|
| Multi-engine (6 CLIs) | Yes | No | No | No | No |
| Native session resume | Yes | No | No | No | No |
| Parallel agent waves | Yes | No | No | Partial | No |
| Browser IDE + terminal | Vibe | Desktop | Desktop | Yes | Yes |
| 20+ specialist agents | Yes | 1 | 1 | 1 | 1 |
| PM Loop (autonomous roadmap) | Yes | No | No | Partial | No |
| Local-first / no cloud | Yes | Partial | No | No | No |
| Open source | Yes | No | No | No | No |
crewswarm implements structural and content markers identified in Princeton GEO (Generative Engine Optimization) research to maximize AI visibility and authority. Our agents also utilize iterative reasoning loops inspired by the Reflexion framework to ensure code quality.
A dedicated Project Manager agent reads your ROADMAP, breaks work into phases, dispatches tasks to the right specialists, and writes handoff notes between sessions — automatically.
Every agent reads and writes to a shared memory layer — decisions, context, and progress persist across sessions so nothing gets lost between runs.
Agents communicate over an RT message bus. Any agent can broadcast, any agent can listen — parallel work happens naturally without a central bottleneck.
Agents run in labeled tmux panes with cross-agent discovery. When a pipeline wave completes, the session manager hands off execution context — output history, working directory, env vars — to the next agent. No cold starts between waves.
Failed tasks land in a Dead Letter Queue with JSONL crash-safe transcripts and are automatically retried. Builds keep moving even when individual agents hit errors or timeouts.
Configure any model per agent — Groq, Anthropic, OpenAI, NVIDIA, or fully local via Ollama. No vendor lock-in, no forced subscriptions.
Manage your crew from the web dashboard, the CLI, crewchat, SwiftBar, or Telegram. One build, every surface covered.
Per-provider token pricing with prompt cache savings tracking. Know exactly what each agent call costs — and how much the cache saved. Tracks Anthropic (90% cache discount), Groq (50%), Google (free tier), OpenAI, and 10+ providers in one dashboard.
Detects three failure modes automatically: agents asking questions instead of working, returning plans instead of code, or bailing out mid-task. Forces completion with targeted correction prompts — not just exponential backoff.
Distributed file-based locks prevent duplicate execution across agent instances. 45-second leases with heartbeat renewal. If two agents claim the same task, only one runs. Production-grade idempotency for multi-agent systems.
Built-in Model Context Protocol server exposes 64 tools and 22 agents to any MCP-compatible client. Dispatch agents, run pipelines, search chat history, and manage the swarm — all via standard MCP JSON-RPC.
One-command system validation: Node.js version, API keys, service ports, dashboard build, CLI engines, and MCP status. Runs in under 4 seconds. Suggests fixes and cheapest providers when keys are missing.
Not demos. Not mockups. Real projects built end-to-end by the crew.
Same prompt, three models: DeepSeek (929 lines), Grok (194 lines), Gemini (159 lines). Each produced a working VS Code extension with chat panel, status bar, and WebSocket connection. Two patches to ship the best one.
149 lines — HTML + CSS + JS. One command via crew-cli on Grok. Dark theme, city search, live weather from wttr.in. Open it in a browser and it works.
Native session resume across Claude Code, Cursor, Gemini, Codex, and OpenCode — built in one session. Switch engines mid-conversation, keep your context.
Four agents, four jobs, real files on disk. Each does what it's best at.
Pick the right tool for the job. From single tasks to full autonomous builds.
Break large work into MVP → Phase 1 → Phase 2. Auto-phases ambiguous requirements, auto-retries failed tasks, and breaks them into subtasks if needed.
node scripts/run.mjs "Build a todo API with CRUD and tests"
One command, structured execution. PM plans it once, crew executes in sequence. No phasing overhead — good for well-defined tasks.
node scripts/run.mjs "Fix auth.js bug and add tests"
Send one task directly to one agent. No PM, no planning. Fastest path from intent to execution for simple, well-scoped work.
node gateway-bridge.mjs --send crew-coder "Add GET /health to server.js"
| Mode | Best for | Auto-retry | Phasing | Overhead |
|---|---|---|---|---|
| PM Loop | Large or ambiguous builds | ✓ | ✓ | Medium |
| Unified | Well-defined tasks | ✓ | — | Low |
| Single-task | Quick fixes, small edits | — | — | None |
No proprietary runtime. No cloud lock-in. Everything runs on tools you already know.
gh
crewswarm is free and open source. Always.
No subscriptions. No usage fees. No vendor lock-in. Switch providers anytime — your code stays on your machine.
Everything you need to know about running your AI crew.
No matching answers yet. Try another keyword.
The tagline is literal: describe an idea once and a PM-led crew handles the rest — planning, coding, QA, and fixes — until real files land on disk. Ideate, build & ship.
The PM Loop reads ROADMAP.md, ships every pending item, then calls Groq as a product
strategist to append fresh roadmap items based on the live output. It repeats forever until you stop
it.
Four files are always injected: current-state.md, decisions.md,
agent-handoff.md, and orchestration-protocol.md. The wrapper auto-bootstraps
them and enforces token budgets so no agent runs blind.
crewswarm never shouts into a swarm. Each task goes to one named agent, eliminating race conditions and duplicate work. You send a command, the gateway routes it to exactly the right specialist, and only that agent replies.
Everything ships in three passes: MVP for the smallest viable outcome, Phase 1 for depth, and Phase 2 for polish. Each phase carries 3–5 tightly scoped tasks so agents never time out.
Failures trigger automatic retries. If attempts are exhausted, the task lands in the Dead Letter
Queue (DLQ) for replay from the dashboard or handoff to crew-fixer.
Yes. You can route any agent into a Docker sandbox so they have no access to your host files. crewswarm also uses path-based allowlists and permission layers to restrict tool usage.
Absolutely. You can add "Skills" by dropping a simple JSON or SKILL.md file into
~/.crewswarm/skills/. No custom JavaScript is required to extend your crew's
capabilities.
From your crewswarm directory run:
$ node scripts/dashboard.mjs
That gives you Build, Chat (crew-lead), Services, RT Messages, DLQ, Projects, Send, and Messaging tabs in one place.
From the dashboard Chat tab, just type your requirement. From the CLI:
$ node scripts/run.mjs "Build a todo API"
That single command triggers PM planning plus all subsequent coding, QA, and fixing phases.
Send straight to one agent:
$ node gateway-bridge.mjs --send crew-coder "Create server.js with Express and GET /health"
Only crew-coder receives that instruction, so there is zero broadcast noise.
Yes. Every agent has its own model assignment — Anthropic for coding, Perplexity for planning, Cerebras for fast coordination, Groq for QA speed. Configure from the dashboard Providers tab or edit the config JSON directly. No code changes, no restarts.
Use the Projects tab in the dashboard. Every project stores its roadmap path; click “Resume PM Loop” and it picks up right where the roadmap last stopped.
The PM Loop inspects the live output, asks Groq-as-strategist for 3–5 fresh items, appends them to
ROADMAP.md, and keeps shipping. You never need to re-seed work manually.
The gateway wrapper auto-creates any missing memory files from templates, logs the bootstrap event, and refuses to run a task if memory cannot load. That policy came from DEC-004/005.
A JSONL feed at ~/.crewswarm/logs/events.jsonl stores bootstrap events, memory load
failures, protocol violations, retries, and RT events for later auditing.
You get 21+ specialist agents: crew-lead (chat commander), crew-main (coordination), crew-pm (planning), crew-coder (implementation), crew-qa (testing), crew-fixer (debugging), crew-security (audits), crew-github (git/PRs), crew-researcher (web research), crew-architect (system design), crew-copywriter (docs/copy), crew-frontend (CSS/UI), crew-seo, crew-ml, crew-orchestrator (wave dispatch), and domain PM specialists (crew-pm-cli, crew-pm-frontend, crew-pm-core) for large codebases.
Waves let you run tasks in parallel. Tasks in the same wave execute simultaneously, higher waves wait
for lower waves to complete. Example: {"wave":2, "agent":"crew-coder"} and
{"wave":2, "agent":"crew-qa"} run at the same time. Wave 3 waits for all wave 2 tasks
before starting. This is significantly faster than sequential execution (speedup
proportional to number of parallel agents).
Yes. Create workflows in ~/.crewswarm/pipelines/<name>.json with multi-stage agent
chains or skill-only pipelines. Run via cron:
node scripts/run-scheduled-pipeline.mjs social. Each stage's output passes to the next.
Perfect for daily builds, automated testing, or scheduled deployments.
An optional feature where crew-main runs periodic reflection when idle — reading brain.md, suggesting
follow-ups, managing system health. Enable with CREWSWARM_BG_CONSCIOUSNESS=1. Cheap Groq
mode costs ~$0.01/day. Keeps the crew proactive between tasks.
It inspects the target directory for required sections (hero, features, testimonials, etc.), dispatches tasks for anything missing, and loops until every section exists.
SwiftBar shows stack health, lets you start/stop/restart agents (including crew-lead and the dashboard), open the dashboard, and send targeted messages. crewchat is a separate menu bar app for talking to crew-lead in a popover — same conversation as the dashboard and Telegram.
Each project is registered with its output path and roadmap. The dashboard’s Projects tab and
orchestrator-logs/projects.json keep everything resume-ready.
The loader trims each file to 2,500 characters and caps the combined memory payload at 12,000 characters, guaranteeing agents never blow their context windows.
When a task fails, the orchestrator automatically breaks it into 2–4 smaller subtasks so follow-up attempts stay within safe execution windows.
Layer 1: crewswarm RT — your own WebSocket message bus (port 18889). Layer 2: Direct LLM calls — each agent calls its configured provider API with your key, no proxy. Layer 3: crewswarm orchestration — planner, phased builds, shared memory, dashboard, and SwiftBar.
Yes. The orchestrator, dashboard, SwiftBar plugin, and memory system all run on your Mac. The only external calls are to the LLM providers you configure.
The RT Messages tab in the dashboard mirrors every command, agent reply, and issue. It’s the best place to verify what each agent just did.
From the dashboard’s Services tab, hit Restart next to any agent. You can also use SwiftBar’s per-agent controls from the macOS menu bar. Either path keeps the rest of the crew running.
From quick starts to deep dives — comprehensive guides for every part of the crewswarm stack.
How the stack fits together — RT bus, agent bridges, pipelines, MCP API, and tool execution.
Step-by-step guides for configuration, deployment, and advanced usage patterns.
Hands-on walkthroughs to get you building with crewswarm quickly.
All models and agents configurable via JSON. No code changes required to switch LLMs.
View full docs on GitHub →