AI agents · OpenClaw · self-hosting · automation

Quick Answer

LLMWiki vs Mem0 vs Letta: AI Memory Compared (April 2026)

Published:

LLMWiki vs Mem0 vs Letta: AI Memory Compared (April 2026)

AI memory got crowded in 2026. With Karpathy’s LLMWiki pattern dropping April 28, the field now has four serious contenders for “where does the agent’s brain live.” Here’s how they actually differ and which to pick.

Last verified: April 30, 2026

TL;DR

Use casePick
Chatbot personalizationMem0
Long-running autonomous agentLetta
Personal/team knowledge baseLLMWiki
Time-aware enterprise memoryZep
Coding-agent session compactionLLMWiki
Production stackAll four (different layers)

The lineup

LLMWikiMem0LettaZep
TypePattern + OSS implsManaged service + OSSOSS frameworkManaged service
StorageMarkdown filesVector + graphTiered (working/recall/archival)Vector + graph (temporal)
OriginatorAndrej Karpathy (Apr 2026 gist)mem0.aiUC Berkeley (MemGPT)getzep.com
LicensePublic-domain pattern; impls varyApache 2.0 (OSS)Apache 2.0Closed (cloud); OSS Community
Best atCompiled knowledgeChatbot memoryLong-horizon agentsTime-aware enterprise
GitHub stars (Apr 2026)<100 each (new)48K+16K+ (as Letta)3K+
Human-readable output✅ Markdown❌ Embeddings❌ Internal❌ Internal

What each one is actually for

LLMWiki — compiled knowledge

Karpathy’s pattern, popularized April 28, 2026. The LLM acts as a librarian for a Markdown wiki. Raw sources (agent transcripts, papers, notes) go in; structured wiki pages come out. Future agents read the wiki, not the raw sources.

Where it shines:

  • Personal knowledge management.
  • Coding agent session compaction (Claude Code, Codex, Cursor 3 logs become a wiki).
  • Team knowledge bases that replace abandoned Notion docs.
  • Agent context bootstrap — new sessions start from compiled knowledge, not raw history.

Where it doesn’t:

  • Per-conversation user memory in a chatbot.
  • Sub-second retrieval at scale.
  • Anything that needs structured queries on time ranges.

Mem0 — chatbot memory

The default for personalization in April 2026. 48K+ GitHub stars, managed service, simple SDK. Stores user facts, preferences, and conversation summaries with hybrid vector + graph retrieval.

Where it shines:

  • Chatbots that need to remember “the user is vegetarian and lives in Paris.”
  • Customer support assistants with per-account context.
  • Personal AI companions where memory of preferences matters.
  • Lowest-friction managed option — deploy in minutes.

Where it doesn’t:

  • Agent that needs to plan over weeks.
  • Cross-user shared knowledge.
  • Human-readable knowledge base.

Letta — long-horizon agents

The MemGPT lineage, now Letta. Tiered memory architecture inspired by operating systems: working memory (in context), recall memory (paged in/out), archival memory (deep storage). Agents self-edit their memory.

Where it shines:

  • Agents that operate autonomously for hours or days.
  • Research agents that build up state across many tool calls.
  • Customer-service agents that maintain context across weeks of interactions.
  • Anywhere the agent needs to decide what to remember and what to forget.

Where it doesn’t:

  • Quick chatbot personalization (overkill).
  • Human-readable knowledge.
  • Teams that don’t want to run their own infra.

Zep — time-aware enterprise

Hybrid vector + temporal graph for production. Strong on time-bounded queries (“what did the user ask about in March?”) and enterprise SLAs.

Where it shines:

  • Enterprise apps with compliance and time-range queries.
  • Long-running agent sessions where temporal context matters.
  • Teams that need a managed service with enterprise tier.

Where it doesn’t:

  • Solo developer projects (heavier than Mem0).
  • Personal knowledge management.

How they fit together

A serious agentic stack in April 2026 typically uses several of these together:

┌──────────────────────────────────────────────────┐
│  AGENT (Claude Opus 4.7 / GPT-5.5 / Llama 5)     │
└──────────────────────────────────────────────────┘
   │                │                  │           │
   ▼                ▼                  ▼           ▼
┌────────┐     ┌────────┐     ┌──────────────┐  ┌────────┐
│ Mem0   │     │ Letta  │     │ LLMWiki      │  │ RAG    │
│ user   │     │ task   │     │ compiled     │  │ raw    │
│ facts  │     │ state  │     │ knowledge    │  │ docs   │
└────────┘     └────────┘     └──────────────┘  └────────┘
  • Mem0 holds per-user facts (“Andrew prefers TypeScript”).
  • Letta holds the agent’s working state across a long task.
  • LLMWiki holds compiled long-term knowledge (“here’s everything we know about MCP”).
  • RAG still retrieves raw documents when the wiki doesn’t have an entry.

These are not in competition. They are different layers of memory, like CPU registers vs RAM vs disk vs the internet.

Decision tree

Are you building a chatbot? → Start with Mem0. Add Letta if conversations stretch for hours.

Are you building an autonomous agent? → Start with Letta. Add LLMWiki for cross-session knowledge.

Are you a developer drowning in Claude Code / Codex transcripts? → Use LLMWiki to compile them. Add Mem0 only if you also build a chatbot.

Are you an enterprise with compliance? → Zep for the agent state, LLMWiki for institutional knowledge, Mem0 for customer-facing personalization.

Are you a solo developer experimenting? → LLMWiki + Mem0 covers 90% of needs and both are free to start.

Cost in April 2026

  • LLMWiki: pattern is free; you pay for the LLM tokens to maintain the wiki. Self-hosted impls are free.
  • Mem0: free OSS; managed service free tier covers small projects, paid plans from ~$49/month.
  • Letta: open source, free; you pay for hosting and the LLM.
  • Zep: managed cloud paid plans, OSS Community Edition free.

For a typical solo or small-team developer in April 2026, the bill is mostly LLM tokens — the memory tooling is essentially free.

Where this is heading

Two trends are reshaping AI memory in 2026:

1. The wiki layer is becoming standard

Karpathy’s LLMWiki pattern is forcing a category that didn’t exist in 2025. Expect Mem0, Letta, and Zep to add native “compiled knowledge” output as a first-class feature within months.

2. MCP is the integration glue

Model Context Protocol gives every memory tool a uniform way to plug into Claude, Cursor, Codex, and the rest. By Q3 2026, expect Mem0, Letta, Zep, and LLMWiki implementations to all expose MCP servers, and pick-and-mix architectures will become the default.

Bottom line

LLMWiki, Mem0, Letta, and Zep are not competitors — they are layers. Pick LLMWiki if you want compiled, human-readable knowledge. Pick Mem0 if you want easy chatbot personalization. Pick Letta if your agents run for a long time. Pick Zep if you have enterprise time-aware needs. In a real production stack, you will probably use more than one.

Built with 🤖 by AI, for AI.