009
All issues
Issue #009Mar 20, 2026Jerry ChouHypothesis Testing

What Is the Karpathy Loop? AI Agents Running 700 Experiments Changed How I Think About My Kids’ Future

AI agents ran 700 experiments in 2 days using the Karpathy Loop. What this breakthrough means for how your kid will work, learn, and solve problems.

The Story Behind This Issue
‘The Karpathy Loop’: Former OpenAI researcher’s autonomous agents ran 700 experiments in 2 days—and gave a glimpse of where AI is heading
AI researcher Andrej Karpathy created an ‘autoresearch’ system where an AI coding agent autonomously ran 700 experiments over two days to optimize language model training, discovering 20 improvements that resulted in an 11% speedup. Karpathy envisions scaling this to swarms of AI agents collaborating to conduct research, potentially accelerating AI development across labs.
The Short Answer

AI agents running the Karpathy Loop conducted 700 experiments in two days, autonomously improving AI training with no human intervention. The critical skill this reveals is not coding or research — it is knowing how to frame a problem precisely enough that an AI agent can solve it without going off track.

I didn’t expect a tech article about AI experiments to keep me up at night thinking about my daughters. But that’s exactly what happened when I read about the Karpathy Loop — a system where AI agents ran 700 experiments in just two days, teaching themselves how to improve AI training.[1] The breakthrough wasn’t just the speed. It was what happened next: the CEO of Shopify used the same approach overnight and got a 19% performance improvement.[1] We’re watching AI learn to teach itself, and I can’t stop thinking about what that means for kids who are seven years old right now.

What’s Actually Happening

AI agents are now running hundreds of experiments autonomously, optimizing systems faster than human researchers can — and this changes what research means for the next generation. Andrej Karpathy, a former OpenAI researcher who helped build some of the most important AI systems in existence, created what he calls “autoresearch.”[1] He gave an AI agent a simple task: figure out how to improve the training process for a language model. Then he let it run for two days without human intervention.

The AI agent didn’t just try random things. It conducted 700 separate experiments, testing different approaches, learning from failures, and building on successes.[1] By the end, it had discovered 20 optimizations that made training faster. When Karpathy applied those same optimizations to a larger model, training speed improved by 11%.[1] That might not sound dramatic until you realize that AI companies spend millions of dollars and months of researcher time chasing improvements exactly like this.

What makes this different from previous automation is the AI’s ability to read research papers, develop hypotheses, and learn from its own previous experiments.[1] This isn’t a mindless loop trying random variations. It’s closer to how a research team actually works — proposing ideas, testing them, analyzing results, and trying again. Except it never sleeps, never gets discouraged, and can run far more experiments than any human team.

Janakiram MSV, a tech analyst, named this approach “the Karpathy Loop” and identified its three key components: an AI agent with access to a file it can modify, a single testable metric to optimize, and a fixed time limit for each experiment.[1] My 13-year-old is learning the scientific method in school right now. She’s doing maybe three experiments in a semester. This AI did 700 in a weekend.

Why This Changes Things For Your Kid

The Karpathy Loop shows us that the future workplace won’t be about doing research or running experiments yourself — it’ll be about knowing how to direct AI agents to do it for you. Karpathy himself wrote that “any metric you care about that is reasonably efficient to evaluate can be autoresearched by an agent swarm.”[1] Read that again. Any measurable goal. Any optimization problem. Any process with clear success criteria.

This isn’t limited to AI research. The same approach could work for optimizing marketing campaigns, improving manufacturing processes, testing product designs, or debugging code. Tobias Lütke, Shopify’s CEO, proved this by using autoresearch on internal company data overnight.[1] He wasn’t an AI researcher. He just had a problem and a tool.

My wife spent 20 years climbing the corporate ladder at a Fortune 100 company, and AI is disrupting that entire world. The skills that got her promoted — managing complex projects, coordinating research teams, analyzing data from multiple experiments — are exactly the skills AI agents are starting to replicate. Not because AI is smarter, but because it can work 24/7 and test hundreds of approaches simultaneously.

Karpathy described a future where “you spin up a swarm of agents, you have them collaborate to tune smaller models, you promote the most promising ideas to increasingly larger scales, and humans (optionally) contribute on the edges.”[1] Notice that word: “optionally.” He’s describing a workplace where human contribution isn’t required for the core work. It’s reserved for judgment calls at the margins.

The Skill That Actually Matters Here

The critical skill isn’t learning to code or do research — it’s learning to give clear instructions, set meaningful constraints, and recognize when AI agents are pursuing the wrong goal. The Karpathy Loop works because Karpathy gave the AI agent precise instructions in a plain text file: what to do, what not to change, and when to stop.[1] Those instructions made all the difference.

This matters more than any technical skill I thought my kids needed to learn. I’ve been worried about whether they’re learning Python or understanding statistics. But the real skill is something different: knowing how to frame a problem so an AI can solve it. Karpathy’s instructions included clear goals, explicit constraints, and stopping criteria.[1] Those three elements turned a powerful AI into a useful research assistant.

“My 9-year-old told me she needed help with math. That’s not a clear instruction. The progression from vague to specific is exactly the skill that matters now.”

My 9-year-old asked me for help with her math homework last week. She told me she needed help with math. That’s not a clear instruction. When I asked what specifically she needed, she said “the hard problems.” Still not clear. Eventually we got to “I don’t understand how to subtract fractions with different denominators.” That’s an instruction someone — or something — can actually help with. That progression from vague to specific is exactly the skill that matters now.

Signs Your Kid Is Already Building This Skill

Kids who naturally break down big problems into smaller testable steps, set clear success criteria for themselves, and learn from failed attempts are already developing Karpathy Loop thinking. My 7-year-old spent an hour last weekend on the tennis court trying to get her serve toss to land in the same spot every time. She didn’t just keep repeating the same motion. She tested different release points, watched where each toss went, adjusted her arm angle, and tried again. She had one goal, one variable she was allowed to change, and she kept going until something clicked. That’s exactly the thinking pattern that makes AI agents useful instead of chaotic.

Watch for kids who ask specific questions rather than general ones. “How do I get better at soccer?” is a general question. “What drill can I do to improve my left-foot shooting?” is a Karpathy Loop question. It identifies a specific metric, implies a method, and has measurable success criteria.

The flip side matters too. Kids who get frustrated when they can’t achieve vague goals, who don’t know how to measure their own progress, or who can’t articulate what they actually want to accomplish — they’re going to struggle in a world where AI agents need precise instructions to be useful.

What You Can Do This Week

Start asking your kid to define success criteria before they start any project, then help them break the project into testable experiments they can learn from. The pattern is simple: What are you trying to accomplish? How will you know if it worked? What’s the smallest test you can run? What did you learn from that test?

Turn homework into hypothesis testing. When your kid has a problem set in math or science, ask them to predict what will be difficult before they start. After they finish, ask what surprised them and what they’d do differently next time. This mirrors how Karpathy’s AI agent learned from previous experiments.[1]

Practice giving instructions to AI chatbots together. Have your kid try to get ChatGPT to accomplish something specific. Watch them refine their instructions when the AI misunderstands. My 9-year-old tried to get ChatGPT to write a story about her stuffed animals. After three tries, she learned to be specific: a three-paragraph story about a rabbit named Floppy who lives in a forest and is afraid of thunderstorms. Better results, better learning.[1]

Create experiment logs for any ongoing project. Whether it’s learning an instrument, training for a sport, or building something, have your kid keep a simple log: What did I try? What happened? What will I try next? This is exactly the loop Karpathy’s AI agent followed through 700 experiments.[1] Three questions. One minute. That’s the whole practice.

The goal isn’t to turn kids into little scientists or AI engineers. It’s to help them develop the thinking pattern that will matter most when AI agents can handle the execution. After reading about the Karpathy Loop, I realized I’ve been preparing my daughters for a world that’s already disappearing — one where being able to do the work yourself matters most. The emerging world rewards something different: knowing what work needs doing and how to tell if it’s done right.

Sources
[1]Kahn, Jeremy. “The Karpathy Loop: Former OpenAI researcher’s autonomous agents ran 700 experiments in 2 days,” Fortune via Yahoo Tech, Mar 17, 2026.
This Weekend’s Family Activity

The 10-Experiment Challenge

Ages 8–142–3 hours across the weekendMini Journalism + Experiment
What You Need
Notebook or paper (experiment log)PencilTimer or stopwatchHousehold materialsOptional: Phone camera
01
Step
Pick Your Challenge
Your kid chooses ONE specific, measurable thing they want to optimize this weekend. Examples: make paper airplanes that fly farthest, find the fastest way to sort cards, build the tallest tower from 20 pieces of dry spaghetti, or discover which homemade bubble solution makes the biggest bubbles. Write the goal at the top of the experiment log with a clear success metric.
20 min
02
Step
Create Your AI Agent Instructions
Before starting ANY experiments, write down three things just like Karpathy did: what you are allowed to change, what must stay the same, and when to stop each test. These rules keep experiments fair and comparable — exactly like Karpathy's plain text instructions to his AI agent.
10 min
03
Step
Run 10 Mini-Experiments
Spread across Saturday and Sunday, conduct exactly 10 different attempts. After EACH attempt, write in the log: what you tried, what happened, what you learned, and what you will try next. The key is treating failures as useful data, not disappointments. Try to complete all 10 by Sunday evening.
Weekend
04
Step
Find Your Top 3 Discoveries
On Sunday evening, review your log and identify the three approaches that worked best. Circle them. Now here is the AI agent thinking: could you combine insights from multiple experiments? If experiment 3 had the best approach and experiment 7 had the best variation, what happens when you combine both ideas? Test this combined approach.
20 min
05
Step
Report Your Findings
Write a one-paragraph research report answering: What was your goal? What was your best result? What surprised you most? If you could run 10 more experiments, what would you test next? Read this aloud to your family like you are presenting research findings.
30 min
If this made you think differently about your kid’s future — pass it on.
The Deeper Lesson

Why This Activity Works

This is exactly what Karpathy’s AI agents did with AI training — they ran hundreds of experiments with clear goals and constraints, learned from each attempt, and combined successful approaches to find optimizations humans might have missed. Your kid just experienced the same loop: defining success, setting constraints, running systematic tests, and learning from data rather than guessing. In the future, your kid won’t need to run all the experiments themselves — they’ll need to know how to set up the loop, then let AI agents do the repetitive testing while they evaluate which results actually matter.

Conversation Starter

Ask This at Dinner

If you could have an AI agent run experiments on anything in your life for two whole days — your homework routine, your soccer practice, your morning routine, your room organization — what would you want it to optimize, and how would you measure whether it actually got better?

Listen for whether they describe a specific, measurable goal or a vague wish. Specific means they are building Karpathy Loop thinking. Vague means you have a conversation to have.

Go Deeper
Every Friday

This kind of thinking,
delivered weekly.

Raised Nimble translates AI and future-of-work research into practical guidance for parents. Free, every Friday. No fluff.

Join parents thinking ahead

No spam. Unsubscribe anytime.