How Anthropic Learned Mythos Was Too Dangerous for Release

Source: bloomberg.com

TL;DR

The story at a glance

Anthropic decided not to release its advanced AI model Mythos publicly after internal tests revealed its exceptional ability to find and exploit software vulnerabilities in critical systems. AI researcher Nicholas Carlini, testing from Bali during a wedding, quickly uncovered techniques to infiltrate widely used computing infrastructure, prompting warnings from the company's experts. Banks, governments like the US Treasury under Secretary Scott Bessent, and partners in Project Glasswing—including Amazon, Apple, Google, Microsoft, and others—are now racing to assess and mitigate risks from this new era of AI-driven threats. This comes amid Anthropic's April 7 announcement of Mythos Preview and Project Glasswing.[[2]](https://www.anthropic.com/glasswing)[[1]](https://www.bloomberg.com/news/features/2026-04-16/how-anthropic-discovered-mythos-ai-was-too-dangerous-for-release)

Key points

Details and context

Mythos, part of Anthropic's Claude series, marks a shift where AI can outperform most humans in coding tasks like vulnerability hunting, chaining exploits autonomously—far beyond prior models like Claude Opus 4.6, which succeeded only twice on tough Firefox tests vs. Mythos's 181 times. Even non-experts can use it overnight for attacks, raising fears of rapid proliferation to bad actors.[[2]](https://www.anthropic.com/glasswing)

Anthropic's system card details mitigations like probe classifiers for misuse monitoring (e.g., worms, exploits) but notes restricted access since general release risks outweigh benefits now. The model saturated many benchmarks, prompting Responsible Scaling Policy updates.[[4]](https://www.anthropic.com/claude-mythos-preview-system-card)

This follows a data leak revealing Mythos in March; responses include global regulator meetings, with some questioning hype but most affirming the capability jump per UK's AISI evaluations.[[5]](https://fortune.com/2026/03/26/anthropic-says-testing-mythos-powerful-new-ai-model-after-data-leak-reveals-its-existence-step-change-in-capabilities)

Key quotes

"I've found more bugs in the last few weeks with Claude Mythos than in the rest of my entire life combined." — Nicholas Carlini, AI researcher affiliated with Anthropic and Google DeepMind.[[6]](https://mashable.com/article/claude-mythos-preview-project-glasswing-pr-stunt-cybersecurity-experts)

"Dangerous stuff." — Dario Amodei, Anthropic CEO.[[3]](https://www.bloomberg.com/opinion/articles/2026-04-15/anthropic-mythos-ai-is-a-wake-up-call-for-everyone-not-just-banks)

Why it matters

AI models like Mythos signal a cybersecurity tipping point where automated attacks could outpace human defenses, threatening economies, infrastructure, and national security if misused. Companies and users face urgent needs to harden software stacks, while investors eye AI safety firms; banks and tech giants must prioritize patching as zero-days surface faster. Watch government access demands, Glasswing patching results in 90 days, and safeguards in Anthropic's next Opus release, though similar capabilities may soon proliferate regardless.[[2]](https://www.anthropic.com/glasswing)