AI agents · OpenClaw · self-hosting · automation

Quick Answer

What is LLMWiki? Karpathy's AI Knowledge Pattern (Apr 2026)

Published:

What is LLMWiki? Karpathy’s AI Knowledge Pattern (April 2026)

The hot AI pattern of late April 2026 isn’t a tool — it’s a workflow. Andrej Karpathy described it as the LLM Wiki pattern in a public gist, and within 48 hours multiple open-source implementations hit GitHub. Here’s what it actually is and how to use it.

Last verified: April 30, 2026

The short answer

LLMWiki is a pattern where:

  1. Raw sources go in — notes, transcripts, agent session logs, web clippings, code snippets.
  2. An LLM continuously maintains a structured Markdown wiki summarizing the sources, indexing concepts, and cross-linking topics.
  3. Future LLM sessions read the wiki first instead of re-reading raw sources.

The wiki is human-readable Markdown. The LLM is the librarian. You can read it. Other LLMs can read it. It grows, gets edited, and becomes the compiled knowledge layer between your raw input firehose and your AI agents.

Why this is different

Not RAG

RAG (retrieval-augmented generation) chunks documents, embeds them, and retrieves the closest matches at query time. The LLM never restructures the source material — it just searches it.

LLMWiki rewrites the material into a structured Markdown wiki. The wiki has its own organization: topics/, people/, projects/, decisions/. Cross-links connect related entries. The LLM does the organizing work once and amortizes it across all future queries.

Not Notion AI

Notion AI is a feature that summarizes a page on demand. LLMWiki is the pattern of letting the LLM own the entire knowledge base, with a fixed schema (Markdown files in folders) and a continuous compaction loop.

Not just memory

Memory frameworks like Mem0 or Letta store conversation context. LLMWiki stores the distilled output of work — like a research notebook a postdoc maintains for their advisor.

The mechanics (from Karpathy’s gist)

The pattern, simplified:

sources/                # raw input
  meeting-2026-04-28.md
  paper-on-mcp.pdf
  cursor-session-3829.txt
  slack-export-week-17/

wiki/                   # LLM-maintained
  index.md              # entry point, table of contents
  topics/
    mcp.md
    cursor-3.md
    nemotron.md
  people/
    karpathy.md
  decisions/
    2026-04-15-switched-to-claude-code.md

A nightly job (or an LLM agent) runs:

  1. Look at new files in sources/.
  2. Decide which existing wiki entries to update and which new entries to create.
  3. Cross-link entries with [[wiki-style links]].
  4. Write a brief changelog of what was added.

You read the wiki yourself. Other agents read the wiki when they need context. The raw sources stay around but are rarely consulted directly.

Why it blew up in April 2026

Three forces converged:

1. Coding agents produce too much context to read

Claude Code, Codex, Cursor 3, and Gemini CLI all produce session transcripts that are individually long and collectively unreadable. An agent that ran for 4 hours generates a transcript no one will ever open. LLMWiki turns that transcript into one or two updated wiki pages.

2. MCP made file-system access trivial

Model Context Protocol matured through 2025-2026. Any compliant client (Claude Desktop, Cursor, Codex, Claude Code) can read and write a local folder via an MCP filesystem server. Building an LLM that owns a Markdown folder went from “weekend project” to “compose two MCP servers.”

3. Karpathy named it

On April 28, 2026, Karpathy published a gist articulating the pattern. He didn’t invent it — engineers had been drifting toward similar workflows for months. But naming it crystallized the design space, and the open-source community shipped within 48 hours.

Open-source implementations (as of April 30, 2026)

ProjectWhat it does
karpathy/llm-wiki (gist)The original spec. Read this first.
Pratiyush/llm-wikiTurns Claude Code, Codex CLI, Copilot, Cursor, and Gemini CLI sessions into a wiki. Rebuilds on every push.
lucasastorian/llmwikiUpload documents, connect Claude via MCP, the LLM writes the wiki.
tjiahen/awesome-llm-wikiCurated list of all current LLM Wiki tools, schemas, and implementations.
gbrain + gstack (Saeloun)Karpathy-style private wiki built on Rails + AI workflow.

Expect 10x more implementations within weeks.

How to build your own (April 2026)

A minimal working LLMWiki, in three pieces:

1. Pick a folder structure

~/wiki/
  index.md
  topics/
  people/
  decisions/
  meta/
    schema.md     # how the wiki is organized
    log.md        # changelog of LLM edits

2. Pick an LLM with file-system MCP

Any of these work in April 2026:

  • Claude Desktop or Claude Code + filesystem MCP server.
  • Cursor 3 Agents Window with /worktree to isolate writes.
  • Codex CLI with the fs plugin.
  • Local model (Llama 5, Qwen 3.6) via Ollama + a custom MCP filesystem.

3. Write a “wiki maintainer” prompt

Something like:

You maintain a Markdown wiki at ~/wiki/. On every run:

  1. Read ~/wiki/meta/schema.md to understand the structure.
  2. Look at new files in ~/inbox/ (raw sources).
  3. Update or create wiki pages, using [[wiki links]] for cross-references.
  4. Append a one-line changelog to ~/wiki/meta/log.md.
  5. Move processed files to ~/inbox/processed/. Be conservative: never delete a wiki page, never overwrite without diff. Use topics/ for concepts, people/ for individuals, decisions/ for choices and rationale.

Run it nightly via cron, or on-demand via your editor.

What it’s good for

  • Coding agent session compaction — turn 4-hour Claude Code transcripts into a one-page summary.
  • Research notes — papers and blog posts go in raw, structured topics come out.
  • Personal knowledge management — the postdoc-notebook use case Karpathy described.
  • Team knowledge bases — replace abandoned Notion wikis with an LLM-maintained one.
  • Agent context bootstrap — point new agents at your wiki on session start, skip the “read 50 files first” warmup.

What it’s not good for

  • Live collaboration — wiki pages are slow-changing; concurrent edits get messy.
  • Highly structured data — invoices, transactions, tables. Use a database.
  • Anything regulatory — LLM-edited prose is not a system of record.
  • Fast retrieval at scale — past ~10K pages, you still want vector search on top.

How LLMWiki relates to existing memory tools

ToolRole
Mem0 / Zep / LettaPer-conversation agent memory
RAGRetrieval over raw documents at query time
Notion AIOn-demand summarization of pages
LLMWikiPersistent, LLM-compiled, human-readable knowledge layer

These are complementary. A serious agent stack in 2026 uses memory (Mem0/Letta) for conversation state, RAG for raw document search, and LLMWiki for compiled long-term knowledge.

Bottom line

LLMWiki is the most important AI workflow concept named in April 2026. It is not a product, it is not magic — it’s the obvious right thing to do once your LLM can read and write a folder. If you run AI agents that produce more text than you can read, start with the karpathy/llm-wiki gist this week.

Built with 🤖 by AI, for AI.