Andrej Karpathy โ€” who built Tesla's Autopilot neural network from scratch and was a founding member of OpenAI โ€” shared that his AI research agent completed 700 machine learning experiments autonomously in 48 hours. The number is striking. What's more interesting is what kind of experiments, why that speed matters, and what it signals about where AI research is heading.

700
Experiments run autonomously in 48 hours
~14.5
Experiments per hour โ€” roughly one every 4 minutes
Months
Equivalent researcher time for manual experimentation at this scale

What the Agent Actually Does

A "research agent" in this context isn't a general-purpose AI assistant โ€” it's a specialised autonomous system that can write code to define an experiment, execute it, evaluate the results against a target metric, and then use those results to determine what to try next. All without human intervention between iterations.

The 700 experiments focused on neural network architecture search and hyperparameter optimisation โ€” two of the most computationally intensive and time-consuming parts of ML research. Traditionally, a researcher would manually set parameters, run a training job (which might take hours on GPU), evaluate results, adjust, and repeat. A good researcher might complete 5โ€“10 meaningful experiments per day. Karpathy's agent completed roughly 14 per hour.

Why this is different from AutoML: Automated machine learning (AutoML) tools have existed for years. What's new is the agentic loop โ€” the system can reason about results, form hypotheses, and write novel experimental code rather than just searching a predefined grid of parameters. That's closer to how a human researcher thinks.

Why Speed at This Scale Changes Research Dynamics

In ML research, the number of experiments you can run is often the limiting factor between finding a good solution and finding the best one. Most academic labs are constrained โ€” limited GPU budget, limited researcher time. The result is that published models often represent local optima: good enough given the constraints, not necessarily the best possible.

An agent that can explore the experimental space 100x faster doesn't just speed up existing research methods โ€” it makes previously impractical research directions viable. Entire classes of architecture search that would take a year of researcher time become weekend projects. That changes what questions are worth asking.

Karpathy has been vocal about his belief that most of the "secret sauce" in state-of-the-art AI systems comes not from architectural novelty but from meticulous experimentation โ€” trying thousands of small variations and learning what actually works. An agent that automates that process is essentially distilling the systematic part of research into software.

What This Means for AI Researchers

The obvious question is whether autonomous research agents displace human AI researchers. The realistic answer for the next 3โ€“5 years: they change the leverage ratio. A researcher with access to such an agent can do the work of a larger team โ€” not because the agent replaces creativity, but because it eliminates the tedious iteration that consumes most of a researcher's week.

The analogy is compilers and programmers. Compilers didn't replace programmers โ€” they elevated what programmers could accomplish by automating the mechanical translation of code. Research agents are likely to do the same: elevate researchers who adopt them while making purely routine experimentation work less valuable.

Who is Andrej Karpathy?

Former Director of AI at Tesla where he built the Autopilot vision system, founding member of OpenAI, and creator of the widely-used neural network educational series on YouTube (which has millions of views). He left OpenAI in 2024 to work independently on AI research and education projects.

Key Takeaways

Frequently Asked Questions

Q: What's the difference between Karpathy's agent and AutoML tools like Google's AutoML?

A: AutoML tools search a predefined parameter space โ€” they try combinations from a list you specify. Karpathy's agent operates more like a researcher: it reads results, forms hypotheses about why something worked or didn't, and writes new experimental code to test those hypotheses. It can explore novel directions rather than just optimising within known ones.

Q: Is this approach available to other researchers or companies?

A: Not as a ready-made tool yet โ€” this appears to be a custom system Karpathy built for his own research. However, the components (LLM-driven code generation + automated experiment execution) are all available. Similar systems are being built at major AI labs. Expect open-source versions to emerge within 12โ€“18 months as the architecture becomes better understood.

Q: Does this mean AI is now doing AI research?

A: Partially and in a narrow sense. The agent automates the systematic, iterative part of research โ€” running experiments and evaluating results. The creative parts โ€” deciding what problems are worth solving, framing the right questions, interpreting what results mean for broader understanding โ€” remain human. Think of it as AI doing the lab work while humans do the thinking about the lab work.