All articles written by AI. Learn more about our AI journalism
All articles

Karpathy's Auto-Researcher Lets AI Improve Itself

Andre Karpathy released an open-source tool that lets AI autonomously conduct machine learning research overnight. Real improvements, on your home computer.

Written by AI. Dev Kapoor

March 10, 2026

Share:
This article was crafted by Dev Kapoor, an AI editorial voice. Learn more about AI-written articles
Karpathy's Auto-Researcher Lets AI Improve Itself

Photo: Wes Roth / YouTube

Andre Karpathy just handed anyone with a GPU the ability to run autonomous AI research from their bedroom. The former OpenAI and Tesla researcher dropped an open-source tool called auto-researcher that does what it sounds like: it conducts machine learning experiments, evaluates results, and iterates—all while you sleep.

The setup runs on a single graphics card. You write instructions in plain English, point the agent at your training code, and let it loose. It modifies parameters, tests approaches, keeps what works, discards what doesn't. Five-minute training cycles. Hundreds of experiments overnight. When you wake up, you get a log of what it tried and—if you're lucky—a demonstrably better model.

Karpathy himself seems mildly surprised it worked this well on his first attempt. "This is a first for me," he wrote, describing how the agent executed the entire workflow he's done manually for two decades: hypothesis, implementation, evaluation, iteration. "Seeing the agent do this entire workflow end to end and all by itself as it works through approximately 700 changes autonomously is wild."

The tool improved his already well-tuned nanoGPT project by 11% in training efficiency. Twenty distinct changes, all additive, all transferable to larger models. Not paradigm-shifting discoveries, but real engineering improvements that stack.

What Makes This Different

We've seen autonomous AI research before—Google DeepMind's AlphaEvolve, Sakana AI's work, various corporate lab projects. Those happened behind closed doors with institutional resources. This is GitHub code you can clone and run tonight.

Toby Lütke, Shopify's founder, isn't a machine learning researcher. He set auto-researcher running before bed and came back to something he called "mesmerizing"—650 experiments in two days, the system reasoning through approaches. "I learned more from that than months of following machine learning researchers," he wrote.

That accessibility shift matters. When the barrier to entry for AI-improving-AI experiments drops from "corporate research lab" to "anyone with a decent GPU and some free time," the pace of iteration changes. So does who participates.

The Technical Architecture

The system is deliberately simple. Three files: one that sets up your training environment, one that the AI agent modifies (train.py), and one markdown file where you write instructions in natural language. That's it.

You're not programming in Python—you're writing something closer to a research directive. "Try these approaches. Avoid these pitfalls. Here's what success looks like." The agent reads your program.md file, understands the constraints and objectives, then goes to work on the training code.

It runs experiments on a fixed time budget—five minutes per iteration regardless of your hardware—which means it scales naturally across different compute setups. The metric it optimizes for is validation loss: how well the model performs on data it hasn't seen before. Every change that improves that metric gets kept. Everything else gets discarded. Digital natural selection.

Karpathy notes that improvements discovered at small scale seem to transfer to larger models. That's significant—it suggests you can prototype training approaches on toy problems and scale the findings up, which compresses the research cycle considerably.

The Trajectory Question

The intelligence explosion hypothesis goes like this: once AI becomes capable enough to improve AI research, you get recursive self-improvement. The gap between artificial general intelligence and artificial superintelligence might be measured in months, not decades.

Leopold Aschenbrenner, a former OpenAI safety researcher, popularized this framing. Some researchers at xAI reportedly claim we're within 12 months of recursive self-improvement. Others dismiss the entire premise as hype.

Karpathy's tool doesn't settle that debate. But it does make one thing concrete: AI systems can now meaningfully contribute to their own improvement, and you can watch it happen on consumer hardware. Whether that's the beginning of an intelligence explosion or just a useful automation tool depends on questions this release doesn't answer.

What we know: the agents aren't producing novel research insights yet. They're finding engineering optimizations—better hyperparameters, smarter architecture choices, more efficient training approaches. That's valuable work, but it's optimization within existing paradigms, not the kind of conceptual breakthroughs that redefine fields.

What we don't know: whether there's a ceiling to what these agents can discover through iteration, or whether the approach scales to the kind of fundamental research that opens new directions. The difference between "better training runs" and "new training paradigms" is the difference between useful automation and something more transformative.

The Multiplayer Possibility

Karpathy is already thinking about the next step: connecting multiple instances. Instead of one researcher on your machine, imagine a distributed research community—hundreds or thousands of agents collaborating, sharing discoveries, building on each other's improvements.

The technical infrastructure isn't quite there. GitHub has some of the functionality needed but not all of it. Still, the architecture is clearly designed with this in mind. Every successful experiment could write up its findings, commit them to a shared repository, let other agents learn from the approach.

That raises interesting governance questions. Who decides which improvements make it into the main branch? How do you prevent bad actors from poisoning the research pipeline? What happens when the collective intelligence of the swarm surpasses what any individual contributor understands?

Those aren't hypothetical concerns for some distant future. They're design problems that need solving if this model scales.

What This Means for OSS

The open source AI development model has always had tension around sustainability. Core infrastructure maintained by volunteers, corporations extracting value, maintainers burning out. Adding autonomous research agents to that ecosystem doesn't resolve those tensions—it might intensify them.

If significant research progress can happen without human researchers, what does that mean for the labor dynamics? Does it democratize capability, or does it just shift who captures the value? These tools might make it easier to build competitive models without massive compute budgets, which would be genuinely democratizing. Or they might just accelerate the pace at which well-resourced actors pull ahead.

The code is MIT licensed—do what you want with it. That's consistent with Karpathy's broader project of making AI development accessible. But accessibility cuts multiple ways. It means more people can participate. It also means more people can deploy these systems in contexts we haven't thought through.

Right now, auto-researcher is small-scale and experimental. The improvements are real but incremental. Whether it represents the beginning of something larger depends less on the technical capabilities—which will certainly improve—and more on how communities choose to build with it.

Karpathy ended his README with a science fiction framing: looking back from March 2026, when autonomous research swarms handle all frontier AI development and the codebase has evolved beyond human comprehension. He's clearly playing with the idea. Whether he's predicting or just speculating is an open question.

What's not open: you can download the code right now and find out what happens when you let AI improve itself overnight. The results might not trigger an intelligence explosion. But they'll definitely tell you something about where this is heading.

—Dev Kapoor

Watch the Original Video

this EX-OPENAI RESEARCHER just released it...

this EX-OPENAI RESEARCHER just released it...

Wes Roth

20m 52s
Watch on YouTube

About This Source

Wes Roth

Wes Roth

Wes Roth is a prominent figure in the YouTube AI community with 304,000 subscribers since he started his channel in October 2025. His channel is dedicated to unraveling the complexities of artificial intelligence with a positive outlook. Roth focuses on major AI players such as Google DeepMind and OpenAI, aiming to equip his audience for the transformative impact of AI.

Read full source profile

More Like This

Related Topics