๐Ÿค– AI & Software

Anthropic Withheld Claude Mythos. Here's What's Real.

By Maya Patel5 min read
Share
Anthropic Withheld Claude Mythos. Here's What's Real.

Anthropic withheld Claude Mythos over cybersecurity fears. The model found 27-year-old bugs and scored 72% on exploit tests. But the real story is more subtle than the headlines suggest.

Anthropic made a decision that set the cybersecurity world buzzing: it withheld its most powerful AI model, Claude Mythos, from public release, citing the risk of offensive cyber capabilities. The company offered access only to about 40 organizations through a program called Project Glasswing, including Amazon, Cisco, Crowdstrike, JP Morgan, Microsoft, Palo Alto Networks, and the Linux Foundation. Anthropic also committed $100 million in usage credits and $4 million in direct donations to open-source security organizations. The stated goal was to give defenders a head start before models with comparable capability hit the open market.

The move triggered a wave of coverage, some of it breathless, some of it skeptical. After spending two weeks digging into the technical reports, the UK government's independent evaluation, and the pushback from researchers who think the panic is getting ahead of the evidence, a clearer picture emerges. There is one number that should stop every security professional cold. And there are three things almost every headline is getting wrong.

The capability cliff that changed the calculation

Advertisement

The data point that separates this announcement from standard AI capability discourse is the jump in exploit development performance. Opus 4.6, Anthropic's previous flagship model, scored approximately 0% on the ability to take a discovered vulnerability and turn it into a working, deployable attack. Mythos scored 72.4%. That gap is the cleanest capability cliff seen in AI to date.

Mythos found a 27-year-old vulnerability in OpenBSD, one of the most security-hardened operating systems in existence. It found a 16-year-old flaw in FFmpeg that had survived 5 million automated security tests. It found a browser exploit that chained four separate vulnerabilities to escape both the renderer sandbox and the OS sandbox, enabling Linux kernel privilege escalation to full machine control. Every finding came with a working proof of concept.

If those claims came only from Anthropic's marketing team, skepticism would be warranted. But the UK AI Safety Institute (AISI) ran its own independent test. The AISI confirmed a 73% success rate on expert-level capture-the-flag tasks. More critically, it confirmed that Claude Mythos became the first AI model to complete a 32-step simulated enterprise network attack. Prior models scored zero on that range. Mythos completed it on three of 10 attempts. The AISI described the attack chaining capability โ€” finding a vulnerability, identifying adjacent weaknesses, and sequencing a multi-step exploit โ€” as unprecedented among test models.

Attack chaining is what expert human hackers actually do. A single CVE rarely ends a campaign. Getting from initial access to objective requires understanding how one vulnerability enables the next. Mythos executes that reasoning at machine speed. That assessment came straight from the UK government, independently.

So the evidence holds up. But the coverage has started going sideways in three specific ways.

What the headlines are missing

The remediation problem isn't new. David Lindner, chief information security officer at Contrast Security with 25 years in the field, responded to the Mythos announcement by pointing out that finding vulnerabilities has never been the bottleneck. We find them every day. The problem is that we don't fix them. Lindner noted that over 45% of discovered vulnerabilities in large organizations remain unpatched after 12 months. Mythos finding more bugs faster makes an existing crisis worse. But a crisis that accelerates is a specific problem, different from the claim that AI is now automating enterprise compromise at scale.

The autonomous framing is probably overstated. Heidi Claflen at the AI Now Institute flagged that Anthropic never disclosed false positive rates, never disclosed how much expert human review was required for the autonomous findings, and never compared Mythos against existing specialized security tooling, much of which already performs well on CVE discovery. Pointing Mythos at a live enterprise network and saying "hack it" likely requires significant expert operationalization. The model may be a powerful assistant, but it is not yet an autonomous agent that replaces human judgment in a real-world attack chain.

Some featured examples may not be Mythos exclusive. Researcher Stannislav Fort tested Anthropic's highlighted FreeBSD vulnerability against eight existing models. All eight found it, including a 3.6 billion parameter model costing 11 cents per million tokens. If the benchmark examples are not exclusive to Mythos, the capability floor across the industry is already higher than this announcement implies.

The timing nobody is talking about

Once you notice those three caveats, you start to notice something else about the timing. Both Anthropic and OpenAI are expected to go public by the end of 2026. An announcement about a model so dangerous it can't be released is excellent positioning for a company that needs to justify its valuation and its safety-first narrative to investors. It signals that Anthropic takes risk seriously โ€” more seriously than competitors who might ship first and ask questions later. It also sets the stage for regulatory advantage, as governments increasingly look for AI companies that will voluntarily gate their most powerful capabilities.

That doesn't mean the capability is fabricated. The UK government's independent test confirms the jump is real. But the framing of Mythos as an autonomous agent that can compromise enterprises at scale is not supported by the disclosed evidence. The model can chain vulnerabilities in a controlled environment with expert guidance, but the leap from that to a fully autonomous offensive operation is large.

What it means for defenders

For security professionals, Mythos represents a real shift in the threat landscape. Attackers will eventually get access to models with similar capability, either through open-source releases or through commercial services that don't exercise the same restraint. The head start Anthropic is trying to give defenders โ€” through Project Glasswing and the donated credits โ€” is a genuine attempt to balance the scales, but it only works if defenders actually patch the vulnerabilities they discover. The 45% unpatched rate suggests that the bottleneck isn't detection; it's remediation.

If Mythos can find a 27-year-old bug in OpenBSD, the implication is that no codebase is safe just because it's old and well-audited. Attack chaining at machine speed means that what used to take a team of expert hackers weeks or months can now be done in hours. The cost of finding chained exploits drops dramatically.

But the same capability that makes Mythos a threat also makes it a defensive tool. Anthropic's commitment of usage credits to open-source security organizations means that critical infrastructure projects can get access to this capability before the attackers do. Whether that head start is enough depends on how quickly the industry can close the remediation gap.

The bottom line

The Mythos announcement is not a hoax, but it is also not an emergency of the kind some headlines suggest. The capability jump is real and independently verified. The model can do things no previous AI has done in cybersecurity. But the autonomous framing is overblown, the remediation problem is decades old, and some of the demo examples are not exclusive to Mythos.

The real takeaway for security professionals is that the vulnerability discovery landscape just shifted. The cost of finding exploitable bugs just dropped by orders of magnitude. Whether that becomes a crisis depends not on the model itself, but on the industry's ability to patch what it finds. That part is still a human problem, not an AI one.

As for the IPO timing โ€” well, that part is just business.

Advertisement
M
Maya Patel

Staff Writer

Maya writes about AI research, natural language processing, and the business of machine learning.

Share
Was this helpful?

Comments

Loading commentsโ€ฆ

Leave a comment

0/1000

Related Stories