All corrections
Substack April 10, 2026 at 05:19 PM

pauseai.substack.com/p/anthropic-has-just-built-an-ai-that

2 corrections found

1
Claim
The system achieved all of this entirely autonomously, without human direction.
Correction

Anthropic says Mythos operated inside a human-designed testing setup and was prompted by users. It was autonomous after setup, but not “without human direction.”

Full reasoning

Anthropic’s own technical writeup contradicts the absolute wording here. The company says Mythos Preview could identify and exploit vulnerabilities “when directed by a user to do so.” It also describes a human-created evaluation scaffold: researchers launched isolated containers, invoked Claude Code with Mythos, prompted it to find vulnerabilities, assigned different agents to different files, and used a final Mythos agent to review bug reports.

Anthropic does say that many specific bugs were found without human intervention after an initial prompt. But that is materially different from saying the whole process happened “entirely autonomously, without human direction.” The testing setup, prompts, task selection, and validation flow were all provided by humans.

2 sources
2
Claim
More than 90 percent of the company’s code is now produced by AI systems.
Correction

Anthropic has been reported as saying company-wide AI-generated code is between 70% and 90%, not above 90%. About 90% applies specifically to Claude Code’s own codebase.

Full reasoning

This sentence overstates Anthropic’s reported company-wide figure. In Fortune’s January 29, 2026 reporting, Anthropic spokespersons said that company-wide the share of AI-generated code is between 70% and 90%. The same report distinguishes that from Claude Code specifically, where about 90% of the code is written by Claude Code itself.

So the article collapses two different claims into one: it presents a product-specific ~90% figure as if it were a confirmed company-wide “more than 90 percent” figure.

1 source
Model: OPENAI_GPT_5 Prompt: v1.16.0