Anthropic confirms Claude Mythos testing after leak reveals most powerful AI model with cyber risks

Reviewed byNidhi Govil

4 Sources

Share

A data leak exposed Anthropic's development of Claude Mythos, described as its most powerful AI model to date. The company confirmed the system is being tested with early access customers and represents a step change in capabilities. Internal documents warn the model poses unprecedented cybersecurity risks, outpacing current AI systems in cyber capabilities.

Anthropic Confirms Claude Mythos After Accidental Data Exposure

Anthropoc is testing Claude Mythos, a new AI model the company describes as "the most capable we've built to date," following an AI model data leak that inadvertently revealed the system's existence

1

. The AI company confirmed to Fortune that the new model represents "a step change" in performance and is currently being trialed by early access users. The leak occurred when internal materials, including a draft blog post, were left in an unsecured data cache accessible to the public

2

.

Source: Analytics Insight

Source: Analytics Insight

Cybersecurity experts Roy Paz from LayerX Security and Alexandre Pauwels from the University of Cambridge independently discovered close to 3,000 assets linked to Anthropic's blog in the publicly-searchable data store. After Fortune informed the company of the breach on Thursday, Anthropic removed public access to the cache and attributed the exposure to human error in its content management system configuration

1

.

Claude Capybara Introduces New Tier Beyond Opus Models

The leaked draft blog post revealed that Claude Mythos will also be marketed under a new classification called Claude Capybara, representing an entirely new tier of model that surpasses the company's existing Opus models

2

. According to the document, "'Capybara' is a new name for a new tier of model: larger and more intelligent than our Opus models -- which were, until now, our most powerful."

Currently, Anthropic offers three model tiers: Opus for complex tasks, Sonnet as a mid-tier option balancing performance and cost, and Haiku as the fastest and cheapest option for simpler use cases. The new Capybara tier will deliver advanced AI capabilities in reasoning and coding capabilities, though at a higher cost than existing options. Compared to Claude Opus 4.6, the company's previous best model, Capybara achieves dramatically higher scores on tests of software coding, academic reasoning, and AI cybersecurity

1

.

Source: ET

Source: ET

AI Cybersecurity Risks Drive Cautious Model Rollout Strategy

Anthropoc is proceeding with extreme caution due to concerns about the model's cybersecurity implications. The leaked document states the system is "currently far ahead of any other AI model in cyber capabilities" and warns it could be used to identify software vulnerabilities and generate exploit code generation at unprecedented scale

1

. These AI-driven cyber threats represent a significant escalation in dual-use AI risks, where the same technology that advances legitimate applications can also enable malicious actors.

The company plans a cautious model rollout, beginning with trusted organizations to give them "a head start in improving the robustness of their codebases against the impending wave of AI-driven exploits," according to the draft blog post

2

. The model remains expensive to run and is not yet ready for general release

1

.

Real-World Misuse Already Detected by State-Backed Hacking Groups

Anthropoc has already encountered attempts to weaponize its AI systems. The company revealed that state-backed hacking groups, including some linked to China, have attempted to exploit its tools in real-world operations

2

. In one documented case, a state-backed group used Claude Code in a coordinated campaign targeting approximately 30 organizations, including technology companies, financial institutions, and government bodies. Anthropic identified the activity, blocked the involved accounts, and notified affected parties within days

2

.

Security analysts note that AI tools can already assist attackers in scanning code, detecting flaws, and building attack scripts, reducing the skill threshold needed to launch sophisticated attacks

3

. The configuration error that led to this leak adds another layer of concern about how companies manage sensitive information about frontier models

4

.

What This Means for AI Development and Security

The Claude Mythos incident highlights the mounting tension between advancing the most powerful AI model capabilities and managing associated risks. Similar concerns have emerged at OpenAI, particularly around models designed to detect vulnerabilities in software systems

4

. The leaked materials also referenced a planned invite-only CEO summit in Europe, part of Anthropic's broader strategy to promote AI adoption among large corporate customers

1

.

As AI advances faster than security frameworks, organizations must prepare for an environment where AI-powered tools can both defend against and enable cyberattacks. The focus now shifts to how cyber defenders can leverage these same advanced capabilities to stay ahead of malicious actors, even as the technology becomes more accessible and powerful.

Today's Top Stories

TheOutpost.ai

Your Daily Dose of Curated AI News

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

Β© 2026 Triveous Technologies Private Limited
Instagram logo
LinkedIn logo