By using this site, you agree to the Privacy Policy and Terms of Use.
Accept
World of SoftwareWorld of SoftwareWorld of Software
  • News
  • Software
  • Mobile
  • Computing
  • Gaming
  • Videos
  • More
    • Gadget
    • Web Stories
    • Trending
    • Press Release
Search
  • Privacy
  • Terms
  • Advertise
  • Contact
Copyright © All Rights Reserved. World of Software.
Reading: Sixteen Claude Agents Built a C Compiler Without Human Intervention… Almost
Share
Sign In
Notification Show More
Font ResizerAa
World of SoftwareWorld of Software
Font ResizerAa
  • Software
  • Mobile
  • Computing
  • Gadget
  • Gaming
  • Videos
Search
  • News
  • Software
  • Mobile
  • Computing
  • Gaming
  • Videos
  • More
    • Gadget
    • Web Stories
    • Trending
    • Press Release
Have an existing account? Sign In
Follow US
  • Privacy
  • Terms
  • Advertise
  • Contact
Copyright © All Rights Reserved. World of Software.
World of Software > News > Sixteen Claude Agents Built a C Compiler Without Human Intervention… Almost
News

Sixteen Claude Agents Built a C Compiler Without Human Intervention… Almost

News Room
Last updated: 2026/02/14 at 7:44 AM
News Room Published 14 February 2026
Share
Sixteen Claude Agents Built a C Compiler Without Human Intervention… Almost
SHARE

In an effort to probe the limits of autonomous software development Anthropic researcher Nicholas Carlini used sixteen Claude Opus 4.6 AI agents to build a Rust-based C compiler from scratch. Working in parallel on a shared repository, the agents coordinated their changes and ultimately produced a compiler capable of building the Linux 6.9 kernel across x86, ARM, and RISC-V, as well as many other open-source projects.

The agents ran roughly 2,000 sessions without human intervention, incurring about $20,000 in API costs. According to Carlini, this “dramatically expands the scope of what’s achievable with LLM agents”.

While Carlini describes the compiler as an “interesting artifact” in its own right, he stresses that the deeper lessons are about “designing harnesses for long-running autonomous agent teams”, ensuring agents remain on track without human oversight and can make progress in parallel:

If you ask for a solution to a long and complex problem, the model may solve part of it, but eventually it will stop and wait for continued input—a question, a status update, or a request for clarification.

Carlini’s approach consisted in “sticking Claude in a simple loop”, so that the agent keeps working on a given task until it’s perfect, then it immediately moves to the next.

He paired this setup with multiple Claude instances running in parallel, each inside its own Docker container but accessing a shared Git repo. This increased efficiency, allowing Claude to tackle multiple tasks at once, and encouraging agent specialization, with some agents handling documentation, others generated code quality, and so on. To synchronize agents, Carlini relied on a simple lock-based scheme:

Claude takes a “lock” on a task by writing a text file to current_tasks/ […]. If two agents try to claim the same task, git’s synchronization forces the second agent to pick a different one.

Once a task is complete, the agent merges other agents’ changes locally, then pushes its branch, and removes the lock. Carlini says “Claude is smart enough to figure out” merge conflicts on its own.

Most notably, in this setup, Carlini does not use an orchestration agent, preferring to “leave it up to each Claude agent to decide how to act”:

In most cases, Claude picks up the “next most obvious” problem. When stuck on a bug, Claude will often maintain a running doc of failed approaches and remaining tasks.

Carlini enforced a number of key practices to ensure success, including maintaining high-quality tests and continuous integration while preventing Claude from spending too much time on testing; assigning distinct agents to separate projects when they were likely to hit the same bug; and specializing agents as mentioned above.

The possibility that multiple agents encountered the same bug simultaneously, generating distinct fixes that would overwrite each other’s work, was a major problem, particularly evident with the Linux kernel. To address this, Carlini employed GCC as a compiler oracle: each agent used GCC to compile a random subset of the kernel tree while Claude’s compiler handled the remainder, refining its output only on that subset.

After two weeks and approximately $20k in API costs, this effort produced a 100k-line compiler that passes 99% of GCC’s torture test, can compile FFmpeg, Redis, PostgreSQL, QEMU, and runs Doom.

Carlini’s effort ignited a wide online debate, with reactions ranging from positive to skeptical and sparked further discussion on its practical and philosophical impact.

X user @chatgpt21 noted that while this was no small feat, it still required a human engineer to “constantly redesign tests, build CI pipelines when agents broke each other’s work, and create workarounds when all 16 agents got stuck on the same bug”.

On the other hand, @hryz3 emphasized that those agents were trained “on the same code they were asked to reproduce”. More sarcastically, @TomFrankly wrote:

They spent $20k in tokens to spit out code that’s in the training data they scraped?

Microsoft’s Steve Sinofsky further qualified the claim that Claude did in two weeks the work that took human engineers 37 years by pointing out that:

It didn’t take GCC 37 years to be built. In 1987 it fully worked for the language as it existed at the time. Over 37 years it evolved with the language, platforms, libraries, optimization and debugging technology, etc.

@WebReflection uncovered another interesting dimension of the debate asking:

How much OSS contribution was done in the making? ‘cause [there will be] no experts’ code to look at as reference in the future if [not] giving anything back to the sources that made any of this possible.

[@RituWithAI summed up]](https://x.com/RituWithAI/status/2019633413402292291?s=20) the implications this might have on software development roles:

We are entering an era where the primary skill for a 10x developer isn’t their ability to solve a complex bug, but their ability to design the automated testing rigs and feedback loops that allow sixteen parallel instances of a model to solve it for them.

As a final note, Carlini himself hinted at the risks that being able to generate code so easily may pose and at the need for “new strategies to navigate safely” this world.

Sign Up For Daily Newsletter

Be keep up! Get the latest breaking news delivered straight to your inbox.
By signing up, you agree to our Terms of Use and acknowledge the data practices in our Privacy Policy. You may unsubscribe at any time.
Share This Article
Facebook Twitter Email Print
Share
What do you think?
Love0
Sad0
Happy0
Sleepy0
Angry0
Dead0
Wink0
Previous Article Carl Pei inaugurates Nothing’s first India flagship store in Bengaluru Carl Pei inaugurates Nothing’s first India flagship store in Bengaluru
Next Article LSEnet: Mastering Automated Data Grouping in Curved Hyperbolic Space | HackerNoon LSEnet: Mastering Automated Data Grouping in Curved Hyperbolic Space | HackerNoon
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Stay Connected

248.1k Like
69.1k Follow
134k Pin
54.3k Follow

Latest News

The Easiest Way To Stop Juggling Multiple AI Tools: 1min.AI For a Flat
The Easiest Way To Stop Juggling Multiple AI Tools: 1min.AI For a Flat $75
News
If your goal is YouTube monetization, long-form isn’t optional.
If your goal is YouTube monetization, long-form isn’t optional.
Computing
Bitcoin biopic ‘Killing Satoshi’ leans into generative AI
Bitcoin biopic ‘Killing Satoshi’ leans into generative AI
News
XIN Summit media day shines a spotlight on Shenzhen’s tech scene · TechNode
XIN Summit media day shines a spotlight on Shenzhen’s tech scene · TechNode
Computing

You Might also Like

The Easiest Way To Stop Juggling Multiple AI Tools: 1min.AI For a Flat
News

The Easiest Way To Stop Juggling Multiple AI Tools: 1min.AI For a Flat $75

4 Min Read
Bitcoin biopic ‘Killing Satoshi’ leans into generative AI
News

Bitcoin biopic ‘Killing Satoshi’ leans into generative AI

3 Min Read
Extreme Networks bets on agentic AI networking as supply chain uncertainty weighs on margins –  News
News

Extreme Networks bets on agentic AI networking as supply chain uncertainty weighs on margins – News

8 Min Read
My uncanny AI valentines
News

My uncanny AI valentines

15 Min Read
//

World of Software is your one-stop website for the latest tech news and updates, follow us now to get the news that matters to you.

Quick Link

  • Privacy Policy
  • Terms of use
  • Advertise
  • Contact

Topics

  • Computing
  • Software
  • Press Release
  • Trending

Sign Up for Our Newsletter

Subscribe to our newsletter to get our newest articles instantly!

World of SoftwareWorld of Software
Follow US
Copyright © All Rights Reserved. World of Software.
Welcome Back!

Sign in to your account

Lost your password?