Source: Andrej Karpathy Just 10X D Everyone S Claude Code (YouTube walkthrough, April 2026)

A walkthrough of Karpathy’s LLM-wiki pattern as a Claude Code workflow — the same pattern this vault is built on. Take raw sources (PDFs, transcripts, web articles), let Claude Code ingest them into a folder of markdown, query the result like personal RAG without a vector database. The creator runs two parallel vaults (YouTube knowledge graph and a personal second brain) and reports a 95% token-usage drop versus prior context-stuffing.

Key Takeaways

  • No fancy RAG required. Karpathy: “I thought I had to reach for fancy RAG, but the LLM has been pretty good about auto-maintaining index files and brief summaries… it reads all the important related data fairly easily at this small scale.” Markdown + indexes + Claude Code is the entire stack.
  • The original prompt is deliberately vague. Karpathy left the spec under-specified so users could hack it. The video shows two divergent vaults from the same prompt — YouTube auto-organized into subfolders (tools, techniques, concepts, sources, people), the personal second brain stayed flat — both correct, shaped by context Claude inferred at setup.
  • One source produces 5–25 wiki pages. Ingesting one AI-2027 article generated 23 pages (1 source, 6 people, 5 organizations, 1 AI-systems page, technical-alignment + geopolitical concepts, plus an analysis). Claude Code does its own chunking and relationship-building.
  • Hub-and-spoke topology emerges automatically. Recurring entities (Claude Code, Perplexity, n8n, named people, frameworks like WAT) become graph hubs without manual relationship-building.
  • Hot cache is project-specific. A hot.md (~500 words) caches recent context for agent-driven projects — saves crawling multiple pages on every query. The YouTube vault skips it; the personal second brain uses it heavily.
  • Lint replaces re-embedding. Karpathy runs LLM “health checks” to find inconsistent data, impute missing data via web search, surface new article candidates. Maintenance is a lint pass, not re-embedding a database.
  • Scale ceiling is hundreds, not millions. Works at ~100 articles / ~half a million words. Beyond that, traditional semantic search / knowledge graph / LightRAG becomes more cost-effective.

Karpathy’s Specific Techniques

  • Vault = raw/ + wiki/ + CLAUDE.md + index + log. Two folders, four files. The CLAUDE.md is written by Claude Code on first run.
  • Web Clipper → raw/ → “ingest” is the entire daily loop. Change the Obsidian Web Clipper default folder from clippings/ to raw/.
  • Tell Claude the project’s purpose before the first ingest. “Second brain for personal/business” vs. “AI research dump” produces materially different wiki structures from the same prompt.
  • Query from the vault or point another agent at it. The creator’s executive-assistant agent has a wiki path line in its CLAUDE.md telling it to read the second-brain vault — hot cache, then index, then domain subindex, then search — only when the answer isn’t already in context.
  • Wiki vs. semantic-search RAG (table the creator generated inside the vault itself): wiki uses links over similarity, costs only tokens (no embeddings / vector DB / chunking pipeline), maintenance = lint. Semantic RAG wins above hundreds-of-thousands of documents.

Recent Signals

[Reddit signal — r/ClaudeAI 2026-05-09]: r/ClaudeAI post 1t89g1j (“Best Claude.md files for claude code”, score 180) surfaced two artifacts worth tracking: (1) forrestchang/andrej-karpathy-skills — a Karpathy-named skill bundle with a CLAUDE.md framed as “behavioral guidelines to reduce common LLM coding mistakes,” claimed in-thread to be “one of the most famous/fastest growing GitHub repos” (vendor-effort verification still pending); and (2) path-frontmatter rule files — Anthropic’s memory#organize-rules-with-claude/rules/ doc describes splitting CLAUDE.md into multiple surgical rule files loaded on demand when Claude touches a path matching the rule’s frontmatter (sample: milis92/nestjs-boilerplate). Both extend the index-then-progressive-disclosure pattern this article covers.

[Reddit signal — r/ClaudeAI 2026-05-09]: r/ClaudeAI post 1t8aecu (score 206) surfaced Simon Willison’s link to Thariq Shihipar’s “The Unreasonable Effectiveness of HTML when using Claude Code” — Thariq is on the Claude Code team at Anthropic; the thesis is that HTML beats Markdown as an output format because Claude can drop in SVG diagrams, interactive widgets, and in-page navigation. Sample prompt: “Help me review this PR by creating an HTML artifact… color-code findings by severity.” Pairs with this article’s “wiki vs semantic-RAG” framing — links + structured output > raw text dumps.

Try It

  1. Install Obsidian (free), create a new vault, open the folder in Claude Code.
  2. Paste Karpathy’s LLM-wiki gist into Claude Code, followed by: “You are now my LLM wiki agent. Implement this exact idea as my complete second brain. Guide me step by step. Create the CLAUDE.md schema, raw/ and wiki/ folders, and index/log files.”
  3. Before the first ingest, tell Claude what the vault is for (research dump, second brain, knowledge graph) — this shapes the structure.
  4. Install Obsidian Web Clipper. Change its default save folder from clippings/ to raw/.
  5. Drop one source into raw/. Say “ingest this.” Expect 5–25 wiki pages out of one article; first ingest 10–14 minutes.
  6. After 3–4 ingests, ask Claude to lint the wiki — find connections, gaps, inconsistencies.
  7. Add wiki/hot.md only if the vault drives an agent that benefits from “what was I just doing.”

Open Questions

  • The “95% token-usage drop” stat (383 files + 100 transcripts → wiki) cites an unnamed X user. Verify before reusing externally.
  • Karpathy’s exact lint prompt isn’t in the video; the creator paraphrases. This vault’s ## Lint operation is one concrete instantiation; the canonical Karpathy version may differ.
  • No guidance on when to split a vault vs. extend one. The creator runs two (YouTube + personal); the line between “new vault” and “new topic” is judgment.