Anthropic's Claude Mythos leak reveals AI model with unprecedented cybersecurity risks

Reviewed byNidhi Govil

18 Sources

Share

Anthropic confirmed its most powerful AI model yet after an embarrassing data leak exposed details about Claude Mythos. The company warns the model presents unprecedented cybersecurity risks and could enable large-scale cyberattacks that far outpace defenders' capabilities. The leak comes as Anthropic eyes an IPO and faces Pentagon scrutiny.

Anthropic Confirms Powerful New AI Model After Embarrassing Data Leak

Anthropic has confirmed the existence of Claude Mythos, what it describes as a step change in AI performance and "the most capable we've built to date," following an internal data leak that exposed nearly 3,000 unpublished assets

1

. The leak occurred due to human error in the company's content management system, which failed to mark sensitive materials as private, leaving them accessible in a publicly searchable data store

5

. Fortune reporter Bea Nolan first identified the exposed data, which included an unpublished blog post describing the powerful new AI model as "by far the most powerful AI model we've ever developed"

3

.

Source: Digit

Source: Digit

The timing of this security lapse is particularly ironic given that the leaked materials themselves warned about unprecedented cybersecurity risks posed by Claude Mythos

4

. Anthropic's spokesperson confirmed to Fortune that the model is "a general purpose model with meaningful advances in reasoning, coding, and cybersecurity," though it remains in early access with select customers

5

.

Claude Mythos Presents Unprecedented Cybersecurity Risks

The leaked draft blog post revealed that Anthropic is privately warning government officials that Claude Mythos makes large-scale cyberattacks much more likely in 2026

2

. According to the company's own assessment, the model is "currently far ahead of any other AI model in cyber capabilities" and "presages an upcoming wave of models that can exploit vulnerabilities in ways that far outpace the efforts of defenders"

3

.

The model allows AI agents to work autonomously with sophisticated precision to penetrate corporate, government, and municipal systems—essentially a hacker's dream weapon

2

. One source briefed on the coming models told Axios that a large-scale attack could hit this year, with businesses as ripe targets

2

. The threat is amplified by employees testing agents without realizing they're creating entry points for cybercriminals through what the industry calls "shadow AI"

2

.

Source: Axios

Source: Axios

New Capybara Tier and Performance Benchmarks

The leaked materials also revealed a new AI model tier called Capybara, which would sit above Anthropic's current top-tier Opus model

3

. Claude Mythos appears to be the first model in this new tier, with the leaked blog post stating that "Compared to our previous best model, Claude Opus 4.6, Capybara gets dramatically higher scores on tests of software coding, academic reasoning, and cybersecurity, among others"

4

. Currently, Anthropic offers three tier levels: Opus as the most powerful and expensive, followed by Sonnet and Haiku

3

.

The model is reportedly very expensive to train and operate, so expensive that Anthropic isn't sure how to deploy it at this point

1

. The leaked draft indicated that the system's high running costs mean a public launch has not yet been scheduled

5

.

Pentagon Seizes on Leak Amid Ongoing Tensions

The revelation has given ammunition to the Pentagon, which has been locked in conflict with Anthropic after the company refused to allow its models to be used for domestic surveillance or fully autonomous military weapons

1

. Under Secretary of War Emil Michael, who has significant financial ties to Anthropic competitors, posted "Umm…hello? Is it not clear yet that we have a problem here?" following the leak

1

. Michael has been calling CEO Dario Amodei a "liar" with a "god complex" for weeks

1

.

A judge recently handed down a ruling that temporarily blocked the Department of Defense from labeling Anthropic a security risk, though the Pentagon continues fighting in the court of public opinion

1

. Late last year, Anthropic disclosed the first documented case of a cyberattack largely executed by AI—a Chinese state-sponsored group that used AI agents to autonomously hack roughly 30 global targets, with the AI handling 80-90% of tactical operations independently

2

.

Source: Gizmodo

Source: Gizmodo

Implications for Cyber Defenders and Business Security

Anthropic stated in the leaked blog that it was providing early access to organizations to give them "a head start in improving the robustness of their codebases against the impending wave of AI-driven exploits"

3

. A Dark Reading poll found that 48% of cybersecurity professionals now rank agentic AI as the #1 attack vector for 2026—above deepfakes and everything else

2

.

The new models enable bad actors to scale simply with more compute, no longer limited by finite personnel—like "a warehouse full of the most sophisticated criminals who never sleep, learn on the fly and persist until successful—except the warehouse is infinite"

2

. Cybersecurity stocks plunged on Friday following the news

4

.

The leak also exposed details about a planned private summit for European business leaders at a U.K. country manor, which Dario Amodei is set to attend

5

. The timing coincides with reports that Anthropic is eyeing an IPO later this year, making the revelation about this powerful new AI model potentially valuable information for prospective investors

1

.

Today's Top Stories

TheOutpost.ai

Your Daily Dose of Curated AI News

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

© 2026 Triveous Technologies Private Limited
Instagram logo
LinkedIn logo