Source: Andrej Karpathy Just 10X D Everyone S Claude Code (YouTube walkthrough, April 2026)
A walkthrough of Karpathy’s LLM-wiki pattern as a Claude Code workflow — the same pattern this vault is built on. Take raw sources (PDFs, transcripts, web articles), let Claude Code ingest them into a folder of markdown, query the result like personal RAG without a vector database. The creator runs two parallel vaults (YouTube knowledge graph and a personal second brain) and reports a 95% token-usage drop versus prior context-stuffing.
Key Takeaways
- No fancy RAG required. Karpathy: “I thought I had to reach for fancy RAG, but the LLM has been pretty good about auto-maintaining index files and brief summaries… it reads all the important related data fairly easily at this small scale.” Markdown + indexes + Claude Code is the entire stack.
- The original prompt is deliberately vague. Karpathy left the spec under-specified so users could hack it. The video shows two divergent vaults from the same prompt — YouTube auto-organized into subfolders (tools, techniques, concepts, sources, people), the personal second brain stayed flat — both correct, shaped by context Claude inferred at setup.
- One source produces 5–25 wiki pages. Ingesting one AI-2027 article generated 23 pages (1 source, 6 people, 5 organizations, 1 AI-systems page, technical-alignment + geopolitical concepts, plus an analysis). Claude Code does its own chunking and relationship-building.
- Hub-and-spoke topology emerges automatically. Recurring entities (Claude Code, Perplexity, n8n, named people, frameworks like WAT) become graph hubs without manual relationship-building.
- Hot cache is project-specific. A
hot.md(~500 words) caches recent context for agent-driven projects — saves crawling multiple pages on every query. The YouTube vault skips it; the personal second brain uses it heavily. - Lint replaces re-embedding. Karpathy runs LLM “health checks” to find inconsistent data, impute missing data via web search, surface new article candidates. Maintenance is a lint pass, not re-embedding a database.
- Scale ceiling is hundreds, not millions. Works at ~100 articles / ~half a million words. Beyond that, traditional semantic search / knowledge graph / LightRAG becomes more cost-effective.
Karpathy’s Specific Techniques
- Vault =
raw/+wiki/+CLAUDE.md+ index + log. Two folders, four files. TheCLAUDE.mdis written by Claude Code on first run. - Web Clipper →
raw/→ “ingest” is the entire daily loop. Change the Obsidian Web Clipper default folder fromclippings/toraw/. - Tell Claude the project’s purpose before the first ingest. “Second brain for personal/business” vs. “AI research dump” produces materially different wiki structures from the same prompt.
- Query from the vault or point another agent at it. The creator’s executive-assistant agent has a
wiki pathline in itsCLAUDE.mdtelling it to read the second-brain vault — hot cache, then index, then domain subindex, then search — only when the answer isn’t already in context. - Wiki vs. semantic-search RAG (table the creator generated inside the vault itself): wiki uses links over similarity, costs only tokens (no embeddings / vector DB / chunking pipeline), maintenance = lint. Semantic RAG wins above hundreds-of-thousands of documents.
Recent Signals
[Reddit signal — r/ClaudeAI 2026-05-09]: r/ClaudeAI post 1t89g1j (“Best Claude.md files for claude code”, score 180) surfaced two artifacts worth tracking: (1) forrestchang/andrej-karpathy-skills — a Karpathy-named skill bundle with a CLAUDE.md framed as “behavioral guidelines to reduce common LLM coding mistakes,” claimed in-thread to be “one of the most famous/fastest growing GitHub repos” (vendor-effort verification still pending); and (2) path-frontmatter rule files — Anthropic’s memory#organize-rules-with-claude/rules/ doc describes splitting CLAUDE.md into multiple surgical rule files loaded on demand when Claude touches a path matching the rule’s frontmatter (sample: milis92/nestjs-boilerplate). Both extend the index-then-progressive-disclosure pattern this article covers.
[Reddit signal — r/ClaudeAI 2026-05-09]: r/ClaudeAI post 1t8aecu (score 206) surfaced Simon Willison’s link to Thariq Shihipar’s “The Unreasonable Effectiveness of HTML when using Claude Code” — Thariq is on the Claude Code team at Anthropic; the thesis is that HTML beats Markdown as an output format because Claude can drop in SVG diagrams, interactive widgets, and in-page navigation. Sample prompt: “Help me review this PR by creating an HTML artifact… color-code findings by severity.” Pairs with this article’s “wiki vs semantic-RAG” framing — links + structured output > raw text dumps.
Try It
- Install Obsidian (free), create a new vault, open the folder in Claude Code.
- Paste Karpathy’s LLM-wiki gist into Claude Code, followed by: “You are now my LLM wiki agent. Implement this exact idea as my complete second brain. Guide me step by step. Create the CLAUDE.md schema, raw/ and wiki/ folders, and index/log files.”
- Before the first ingest, tell Claude what the vault is for (research dump, second brain, knowledge graph) — this shapes the structure.
- Install Obsidian Web Clipper. Change its default save folder from
clippings/toraw/. - Drop one source into
raw/. Say “ingest this.” Expect 5–25 wiki pages out of one article; first ingest 10–14 minutes. - After 3–4 ingests, ask Claude to lint the wiki — find connections, gaps, inconsistencies.
- Add
wiki/hot.mdonly if the vault drives an agent that benefits from “what was I just doing.”
Related
- Wiki Community Enhancements — survey of 12 GitHub repos building on Karpathy’s pattern, including the hot cache, delta manifest, and contradiction-detection patterns this vault uses.
- Karpathy LLM-Wiki Pattern — community implementations of the original pattern.
- Agent Skills Overview — how skills compose with this kind of project-shaped Claude Code workflow.
- Claude Code Plugins and Marketplaces — packaging the wiki workflow as a distributable skill or plugin.
- Claude Code Subagents — parallel ingest of large raw batches uses one subagent per source file.
- Anti-AI Slop Guide — wiki articles compounding requires resisting generic LLM filler in summaries.
Open Questions
- The “95% token-usage drop” stat (383 files + 100 transcripts → wiki) cites an unnamed X user. Verify before reusing externally.
- Karpathy’s exact lint prompt isn’t in the video; the creator paraphrases. This vault’s
## Lintoperation is one concrete instantiation; the canonical Karpathy version may differ. - No guidance on when to split a vault vs. extend one. The creator runs two (YouTube + personal); the line between “new vault” and “new topic” is judgment.