TL;DR
Multica is an open-source “managed agents” platform from multica-ai — a self-hosted Kanban + chat layer that lets you assign issues to coding agents the same way you assign them to humans. It hit GitHub trending hard this week with +5,421 stars in seven days (20.8K total) on the strength of one simple pitch: “your next 10 hires won’t be human.”
Instead of being yet another agent framework, Multica is the management layer around the agents you already run — Claude Code, OpenAI Codex, OpenClaw, OpenCode, Hermes, Gemini, Pi, and Cursor Agent. You install a small Go daemon on every machine that has those CLIs, point a Next.js frontend at it, and now your agents have profiles, show up on a board, post comments on issues, and report blockers like a teammate.
Key facts:
- Multi-agent, vendor-neutral — works with 8 coding CLIs out of the box (Claude Code, Codex, OpenClaw, OpenCode, Hermes, Gemini, Pi, Cursor Agent)
- Self-hostable — Docker images on GHCR, full
make selfhost-buildpath, or use Multica Cloud - Stack — Next.js 16 frontend, Go backend (Chi + sqlc + WebSockets), PostgreSQL 17 with
pgvector - Real lifecycle — full
enqueue → claim → start → complete/failstate machine with WebSocket progress streaming - Reusable Skills — every solved task can be distilled into a Skill that future agents pull in automatically
- Multi-workspace — workspace-level isolation for teams, roles, and runtimes
- MIT licensed, single
brew install multica-ai/tap/multicafor the CLI on macOS/Linux - Honest limitation: the “Skills” system is great in theory but young — most teams in the repo’s own discussions report they’re still using it the same way they’d use issue templates, not as compounding institutional memory yet
What Problem Does Multica Actually Solve?
If you’ve spent real time pair-programming with Claude Code, Codex, or OpenClaw, you’ve already hit the wall Multica is built around. The agents themselves are great — but everything outside the agent is held together with copy-pasted prompts, half-broken shell scripts, and tmux panes named agent-final-FINAL-2. You re-paste context every morning, babysit one agent at a time, and re-explain your deploy pipeline to every fresh session.
Multica’s bet is that the right abstraction is issues, not prompts. Make agents first-class users in a project tracker. Let humans and agents share the same board, comments, and backlog. Then route work to whichever runtime is free. The README says it plainly: “No more copy-pasting prompts. No more babysitting runs.”
How It Works (Architecture)
┌──────────────┐ ┌──────────────┐ ┌──────────────────┐
│ Next.js │────>│ Go Backend │────>│ PostgreSQL │
│ Frontend │<────│ (Chi + WS) │<────│ (pgvector) │
└──────────────┘ └──────┬───────┘ └──────────────────┘
│
┌──────┴───────┐
│ Agent Daemon │ runs on your machine
└──────────────┘ (Claude Code, Codex, OpenCode,
OpenClaw, Hermes, Gemini,
Pi, Cursor Agent)
Three pieces:
- Frontend — a Next.js 16 (App Router) app that gives you the board, agent profiles, issue detail pages, chat, and skills library.
- Backend — a Go service (Chi router, sqlc, gorilla/websocket) talking to PostgreSQL 17 + pgvector. Vector storage is used for skill retrieval and conversation context.
- Daemon — a small Go binary (
multica daemon) that runs on any machine where you want to actually execute agent work. It auto-detects which CLIs are on yourPATHand registers them as a Runtime.
The Runtime abstraction is the part most reviewers single out as elegant. A runtime is just “a place that can execute agent tasks” — it can be your laptop, a dev server, a CI box, or a cloud instance. Each runtime advertises its capabilities (which CLIs it has, which models are configured), and the backend routes work to whichever one is free and capable.
State transitions for every task go through enqueue → claim → start → complete/fail, and progress streams to the UI over WebSocket so you can watch tail output live.
Quick Install
The fastest path on macOS or Linux:
# Install the CLI
brew install multica-ai/tap/multica
# Configure, log in, and start the local daemon in one shot
multica setup
That gives you a daemon connected to Multica Cloud. If you want to self-host the whole stack:
curl -fsSL https://raw.githubusercontent.com/multica-ai/multica/main/scripts/install.sh \
| bash -s -- --with-server
multica setup self-host
The self-host path pulls official images from GHCR and stands up Postgres, the Go backend, and the Next.js frontend behind Docker. Self-hosting requires Docker; everything else is one binary.
On Windows:
irm https://raw.githubusercontent.com/multica-ai/multica/main/scripts/install.ps1 | iex
Creating Your First Agent
Once the daemon is running and your machine shows up under Settings → Runtimes, agent creation is genuinely just three clicks:
- Go to Settings → Agents → New Agent
- Pick a runtime (your laptop, a dev box, a cloud VM)
- Pick a provider — Claude Code, Codex, OpenClaw, OpenCode, Hermes, Gemini, Pi, or Cursor Agent
- Give it a name like
claude-frontendorcodex-migrations
That name is how the agent shows up on the board, in @-mentions, and in Slack-style comment threads. From there, you create an issue and assign it to the agent, same as a colleague. The daemon claims it, runs the underlying CLI on its runtime, and streams output back to the issue page.
CLI Cheat Sheet
multica login # Browser-based auth
multica daemon start # Start local agent runtime
multica daemon status # Check daemon + runtime health
multica setup # One-shot configure + login + start
multica setup self-host # Same, for self-hosted servers
multica issue list # List issues in your workspace
multica issue create # Create a new issue
multica update # Update CLI to latest version
Most day-to-day work happens in the web UI, but multica issue create is genuinely useful from terminal-side scripts — for example, kicking off “review this PR” tasks from a git pre-push hook.
A Real Example: Multi-Agent Code Review
Here’s a workflow people are actually using in the repo’s discussions: parallel code review with two different models.
# Create a "review this PR" issue and capture the ID
ISSUE_ID=$(multica issue create \
--title "Review PR #847: Add rate limiting middleware" \
--label review \
--body "Diff: https://github.com/acme/api/pull/847" \
--json | jq -r '.id')
# Assign to two agents on different models
multica issue assign "$ISSUE_ID" --agent claude-reviewer
multica issue assign "$ISSUE_ID" --agent codex-reviewer
Both agents pick up the work, run their own review, and post comments back to the same issue thread. You skim both, pick the better feedback, and ship. The issue becomes the durable artifact — diff, two reviews, decisions, and any follow-up tasks all in one place.
The same pattern works for “implement this feature on branch A vs branch B” and “fix this bug, but compare two approaches.” Multi-agent forking is the killer use case nobody quite had a clean UI for before.
Skills: The Compounding-Capability Bet
Multica’s most ambitious feature is Skills — the idea that every solved task should produce a reusable artifact your team’s future agents can pick up. Concretely, a skill is a markdown file plus a small descriptor that gets stored in the database (with vector embeddings via pgvector) and surfaced to relevant agents based on the issue they’re working on.
In practice, a skill ends up looking like a battle-tested runbook:
# Skill: Deploy Astro site to Hetzner via rsync
When to use: any task tagged `deploy` in the `andrew-ooo` workspace.
Steps:
1. Run `npm run build` and verify dist/ exists
2. rsync -avz --delete dist/ root@HOST:/var/www/site/
3. Run health-check.js against the production URL
4. If health check fails, rollback with: ssh HOST 'mv /var/www/site.bak /var/www/site'
When an agent picks up a deploy task, the backend retrieves the most relevant skills via pgvector similarity, injects them into the prompt, and the agent executes against a known-good runbook instead of inventing one from scratch. In theory.
In practice — and this is where the honest review starts — the system is too young for “skills compounding” to be true yet. Most teams are using skills the same way they’d use Notion runbooks: a place to dump tribal knowledge so the next agent doesn’t relearn it. Real automatic distillation of new skills from completed tasks is on the roadmap but not the default flow today.
Multica vs Paperclip vs Roll-Your-Own
The closest comparison is Paperclip, an earlier project in the same “manage your AI workforce” space. The README’s own table sums up the philosophical split:
| Multica | Paperclip | |
|---|---|---|
| Focus | Team AI agent collaboration | Solo AI agent company simulator |
| User model | Multi-user teams + roles | Single board operator |
| Agent interaction | Issues + Chat | Issues + Heartbeat |
| Deployment | Cloud-first | Local-first |
| Management depth | Lightweight (Issues/Projects/Labels) | Heavy governance (Org chart/Approvals/Budgets) |
| Extensibility | Skills system | Skills + Plugins |
Roughly: Multica is Linear-for-AI-teams, Paperclip is SimCity-for-one-AI-CEO. If you have a team of 2–10 humans and want to add agents to the same backlog, Multica fits cleanly. If you’re a solo founder simulating a 30-person org chart with budgets and approvals, Paperclip’s heavier governance is closer to the brief.
The “do nothing” alternative is what most people are already doing: tmux + a folder of prompt.md files + a Discord channel where you paste agent output. That works for one person and one agent. It collapses the moment you want two humans + three agents touching the same project.
Community Reactions
Reception has been notably positive for an OSS launch. From the r/LocalLLM thread “Open-source alternative to Claude’s managed agents… but you run it yourself”:
“Not saying it replaces anything, but it’s an interesting direction if you’ve seen what Claude Managed Agents is trying to do and wanted more control over it. And it works with Claude Code, OpenAI Codex, OpenClaw, and OpenCode.”
From a separate r/LocalLLM thread on agent orchestration on a 128GB Mac:
“I tried paperclip but I personally find it too bulky. I tried this one today: github.com/multica-ai/multica.”
The most thoughtful critique comes from the repo itself — a long-form issue titled “Discussion: Multica still manages AI the way it manages people” (issue #815):
“First, I want to say this clearly: I think Multica is impressive work. A lot of the product is thoughtful and well executed: the runtime abstraction is clean, the agent model is well packaged, the issue/comment/skills/transcript/chat layers all hang together. But the underlying mental model is still ‘a Jira board where some assignees happen to be bots.’ Agents have radically different failure modes than humans.”
That’s a fair shot. The current Multica UX is deeply borrowed from human project tools — Kanban columns, comments, mentions. Whether that abstraction holds at 100 agents per workspace, or whether agents actually need their own native primitives (heartbeats, budgets, approval gates) is the open question of the category.
Honest Limitations
A few things to know before you adopt:
1. Skills don’t compound automatically yet. The pitch — “every solution becomes a reusable skill” — is real architecturally, but the auto-distillation pipeline is still in active development. Today, skills are mostly hand-written runbooks. If you came for the magic, you’ll need to add your own discipline around writing them.
2. PostgreSQL 17 + pgvector is a real dependency. Self-hosting Multica means running and maintaining a Postgres instance. If you don’t already have Postgres ops in your toolkit, the operational surface is non-trivial — especially as the database grows with conversation transcripts and embeddings.
3. Agent failure modes look ugly in a Kanban UI. A human moves a card from “In Progress” to “Blocked” once. An agent stuck in a loop thrashes through five state transitions in 90 seconds and can swamp the board.
4. No built-in budget/spend caps per agent. Each agent CLI handles its own model billing. Multica doesn’t surface “you spent $42 on Claude API in this issue” the way Paperclip does. If cost control matters, you’ll wire that up via your provider’s dashboards.
5. Cloud vs self-hosted feature parity is moving. Multica Cloud gets new features first; the self-hosted GHCR images sometimes lag by a release. Plan to track release notes if you self-host.
Who Should Use Multica?
Good fit:
- Teams of 2–10 already running coding agents and tired of tmux
- Anyone running multiple agent CLIs (Claude Code + Codex, or OpenClaw + OpenCode) who wants one dashboard
- Open-source-leaning shops that don’t want to send all their issues through a SaaS like Linear
- Companies experimenting with parallel code review using two different model families
Bad fit:
- Solo developers who only run one agent — the UI overhead won’t pay back
- Teams that need heavy governance, approval chains, or per-agent budget caps today
- Anyone unwilling to operate Postgres in production (use Multica Cloud instead)
- Shops with a strict “no JavaScript on production servers” policy — the frontend is Next.js
FAQ
What agents does Multica work with?
Out of the box: Claude Code, OpenAI Codex, OpenClaw, OpenCode, Hermes Agent, Gemini CLI, Pi, and Cursor Agent. The daemon auto-detects whichever CLIs are on your PATH. Adding new agents is a relatively small Go change in the daemon’s runtime registry.
Can I self-host Multica without using Multica Cloud at all?
Yes. Run curl -fsSL https://raw.githubusercontent.com/multica-ai/multica/main/scripts/install.sh | bash -s -- --with-server, then multica setup self-host. You get the full stack — frontend, backend, Postgres, daemon — on your own infrastructure. Docker is the only hard prereq. The Cloud tier is convenient but optional.
Is Multica an agent framework or a management layer?
A management layer. It does not implement an agent loop, prompt the model, or handle tool calls itself. It shells out to existing agent CLIs (Claude Code, Codex, etc.) and gives them a Kanban-shaped harness. That separation is deliberate — the agent ecosystem is moving fast, and Multica wants to be the layer that survives whichever framework wins.
How is this different from Paperclip?
Paperclip is a single-operator “AI company simulator” with heavy governance — org charts, approval chains, per-agent budgets. Multica is a multi-user team platform with lighter primitives — issues, comments, labels, skills. If you’re a solo founder simulating a 30-person org, Paperclip fits. If you’re a real 5-person team adding agents to the same backlog, Multica fits.
Does it work with local LLMs / Ollama / open-weight models?
Indirectly. Multica drives agent CLIs, so whatever models those CLIs support, Multica supports. OpenCode, OpenClaw, and Hermes all run against local endpoints. Claude Code and Cursor Agent are tied to their respective hosted APIs. If running everything on a 128GB Mac is your goal, the r/LocalLLM thread linked above has working configs.
How much does it cost?
The OSS project is MIT-licensed and free forever. Multica Cloud has paid tiers for hosted backend + Postgres, but pricing isn’t the focus of this review — check multica.ai for current numbers. Self-hosting costs you whatever your Postgres + small Go service + Next.js frontend burns on your infra (usually a few dollars a month on a single VPS).
Bottom Line
Multica is the cleanest take I’ve seen on the “Linear for AI teammates” idea. The runtime abstraction is genuinely well-designed, the multi-CLI support is unusually broad, and the choice to be a management layer rather than yet another agent framework is the right strategic call. The product is young — Skills don’t compound yet, governance is light, agent-native primitives haven’t crystallized — but it’s young in the way a product that will keep shipping is young, not in the way an abandoned demo is young.
If you have more than one human running more than one coding agent on the same project, Multica is worth an afternoon. The install path is a single command and the worst case is you keep using tmux.
Repo: github.com/multica-ai/multica · 20.8K ⭐ · MIT · 5,421 stars this week