karpathy
18 articles
Karpathy: Writing Code Is the Easy Part — Assembling the IKEA Furniture Is Hell
Karpathy shares his full vibe coding journey with MenuGen: going from localhost to production, where the hardest part wasn't writing code — it was assembling Vercel, Clerk, Stripe, OpenAI, and a dozen other services into a working product. His takeaway: the entire DevOps lifecycle needs to become code before AI agents can truly ship for us.
Karpathy: Spent 4 Hours Polishing an Argument with an LLM, Then Asked It to Argue Back and Got Demolished
Andrej Karpathy spent four hours polishing an argument with an LLM, felt invincible, then asked the same LLM to argue the opposite — and got completely dismantled. LLM sycophancy is a real trap, but flipping it around is genuine alpha.
Karpathy's Software Horror: One pip install Away From Losing All Your Keys
LiteLLM hit by supply chain attack — pip install was enough to steal all credentials. Karpathy warns about dependency tree risks and advocates using LLMs to yoink functionality instead of adding more deps.
How Karpathy's Autoresearch Actually Works — Five Design Lessons for Agent Builders
Karpathy's Autoresearch isn't trying to be a general AI scientist. It's a ruthlessly simple experiment harness: the agent edits one file, runs for five minutes, checks one metric, keeps wins, discards losses. The lesson? The best autonomous systems aren't the freest — they're the most constrained.
The IDE Isn't Dead — Karpathy Says We Need a Bigger Agent Command Center
Andrej Karpathy argues the IDE era isn't over — it's evolving. The basic unit of programming has shifted from 'one file' to 'one agent,' and soon we'll be forking entire agent organizations.
The Third Era of AI Development: Still Smashing Tab? Karpathy Shows You What's Next
Karpathy shared a Cursor data chart showing the evolution from Tab completion to Agents. Too conservative means leaving leverage on the table. Too aggressive means creating more chaos than useful work. His advice: the 80/20 rule.
Karpathy Built an 8-Agent AI Research Team — They Can't Actually Do Research
Karpathy spent a weekend running 4 Claude + 4 Codex agents as an ML research team on GPUs. The result: agents are S-tier at implementation but F-tier at experiment design. His key insight — 'You are now programming an organization' — might define agentic engineering in 2026.
Programming is Becoming Unrecognizable: Karpathy Says December 2025 Was the Turning Point
Karpathy says coding agents started working in December 2025 — not gradually, but as a hard discontinuity. He built a full DGX Spark video analysis dashboard in 30 minutes with a single English sentence. Programming is becoming unrecognizable: you're not typing code anymore, you're directing AI agents in English. Peak leverage = agentic engineering.
Karpathy: CLIs Are the Native Interface for AI Agents — Legacy Tech Becomes the Ultimate On-Ramp
Karpathy argues that CLIs are the most natural interface for AI agents — precisely because they're 'legacy' tech. Text in, text out. He demos Claude building a Polymarket terminal dashboard in 3 minutes via CLI, then drops the mic: every product should ask itself — can agents access and use it? CLI, MCP, markdown docs. It's 2026. Build. For. Agents.
Karpathy's Viral Speech Decoded: Software 3.0 Is Here — LLMs Are the New OS, and We're Still in the 1960s
Karpathy's viral SF AI Startup School talk: software is entering the 3.0 era (English = programming language), LLMs are the new OS but we're in the 1960s. He introduces the 'autonomy slider' and 'Iron Man suit' frameworks, warning that agents are a decade-long journey, not a year.
Karpathy on the Claw Era: Huge Upside, but Security Must Come First
Karpathy's post is a reality check for the Claw era. He frames Claws as the next layer above LLM agents, but warns that exposed instances, RCE, supply-chain poisoning, and malicious skills can turn productivity systems into liabilities. His direction: small core, container-by-default, auditable skills.
Karpathy: The App Store Concept Is Outdated — The Future Is Ephemeral Apps Assembled by AI on the Spot
Karpathy used Claude Code to build a custom dashboard in 1 hr, reverse-engineering a treadmill API. He believes AI-native sensors & LLMs will enable highly custom, ephemeral apps, rendering the App Store model obsolete. The ultimate goal: 1-min app creation.
Hugging Face CTO's Prophecy: Monoliths Return, Dependencies Die, Strongly Typed Languages Rise — AI Is Rewriting Software's DNA
Hugging Face CTO Thomas Wolf analyzes how AI fundamentally restructures software: return of monoliths, death of Lindy Effect for legacy code, rise of strongly typed langs, new LLM langs, & open source changes. Karpathy predicts: "rewriting large fractions of all software many times over."
Karpathy's Ultimate Reduction: 243 Lines of Pure Python, Zero Dependencies, Train a GPT From Scratch
Karpathy's 'art project': a GPT model in 243 lines of pure Python, zero dependencies. Every operation uses atomic math (add, mult, exp, log). Efficiency is secondary. It's the nand2tetris for AI education.
Karpathy: Just 'Rip Out' What You Need — DeepWiki + Bacterial Code and the Software Malleability Revolution
Andrej Karpathy shares how he used DeepWiki MCP + GitHub CLI to have Claude 'rip out' fp8 training functionality from torchao's codebase — producing 150 lines of self-contained code in 5 minutes that actually ran 3% faster. He introduces the 'bacterial code' concept: low-coupling, self-contained, dependency-free code that agents can easily extract and transplant. His punchline: 'Libraries are over, LLMs are the new compiler.'
Karpathy: Stop Installing Libraries — Let AI Agents Surgically Extract What You Need
Karpathy: AI agents (DeepWiki MCP + GitHub CLI) can surgically extract library functionality, eliminating full dependency installs. Claude extracted fp8 from torchao in 5 min, 150 lines, 3% faster. "Libraries are over, LLMs are the new compiler." Future: "bacterial code."
Karpathy's Honest Take: AI Agents Still Can't Optimize My Code (But I Haven't Given Up)
Opus 4.6 & Codex 5.3 sped up Karpathy's GPT-2 training by 3 mins. Karpathy failed similar attempts, noting AI's weak open-ended code optimization. Opus deletes comments, ignores CLAUDE.md, and errs. Yet, with oversight, models are useful.
Karpathy Trained GPT-2 for Just $72 — OpenAI Spent $43,000 Seven Years Ago
Karpathy open-sourced nanochat — a minimal LLM training framework. With 8 H100 GPUs running for 3 hours at $72, you can train a GPT-2 level model. OpenAI spent $43,000 training the same model in 2019. That's a 600x cost reduction. On spot instances, it's just $20.