Anthropic's Mythos AI faces unauthorized access as security breach exposes controlled rollout flaws

Reviewed byNidhi Govil

37 Sources

Share

Anthropic is investigating reports that unauthorized users gained access to Mythos, its restricted cybersecurity AI model, through a third-party vendor on the day of its announcement. The incident highlights vulnerabilities in controlled AI releases and raises questions about the model's actual capabilities versus the company's dramatic safety messaging.

Anthropic Investigates Unauthorized Access to Mythos AI Model

Anthropic is investigating a security breach involving Mythos, its highly restricted cybersecurity AI model designed exclusively for enterprise security applications. According to Bloomberg, a group of unauthorized users gained access to the AI model through a third-party vendor environment on the same day Anthropic publicly announced Project Glasswing, the controlled release program

1

. The company confirmed it is examining claims of unauthorized access but stated that no evidence suggests the activity impacted Anthropic's core systems

4

.

Source: Market Screener

Source: Market Screener

An Anthropic spokesperson told media outlets: "We're investigating a report claiming unauthorized access to Claude Mythos Preview through one of our third-party vendor environments"

5

. The incident represents an awkward turn for a company that has built its brand on taking AI safety seriously while positioning Mythos as too dangerous for public release.

How Unauthorized Users Accessed the Restricted Hacking Tool

The unauthorized group accessed Mythos through a combination of insider knowledge and predictable security patterns rather than sophisticated technical exploits. Members of a private Discord channel made "an educated guess about the model's online location based on knowledge about the format Anthropic has used for other models," Bloomberg reported

1

. The group also leveraged access through a contractor who works for a third-party vendor evaluating Anthropic models.

Source: New York Post

Source: New York Post

Details about Anthropic's model hosting patterns were reportedly exposed in the recent Mercor data breach. Mercor, an AI staffing startup that supplies specialized contractors to major AI labs including Anthropic, was affected by a LiteLLM supply-chain attack earlier this month

5

. Security researcher Lukasz Olejnik described this type of failure as "entirely imaginable" and something the cybersecurity industry has routinely dealt with for the last 20 years

3

.

The group provided evidence to Bloomberg in the form of screenshots and a live demonstration of the software, claiming they were "interested in playing around with new models, not wreaking havoc with them"

1

.

Project Glasswing and the Controlled Release Strategy

Mythos was released to a select number of vendors through Project Glasswing, including major technology companies like Apple, Amazon Web Services, Google, Microsoft, NVIDIA, and financial institutions like JPMorganChase

2

. The limited release aimed to prevent weaponization by bad actors while allowing organizations to identify cybersecurity vulnerabilities in their own systems before criminals could exploit them.

Source: Mashable

Source: Mashable

According to Anthropic, Mythos can discover flaws in virtually any software and allegedly found "thousands of high- and critical-severity vulnerabilities" in operating systems and other software

2

. The company warned that "the fallout -- for economies, public safety, and national security -- could be severe" if the model fell into the wrong hands.

Mozilla reported that Mythos helped its team find 271 vulnerabilities in Firefox 150, though Mozilla CTO Bobby Holley noted that "we also haven't seen any bugs that couldn't have been found by an elite human researcher"

2

. This assessment suggests the AI model excels at speed and thoroughness rather than discovering entirely novel vulnerability types.

Supply-Chain Threats Expose Weaknesses in AI Security

The breach highlights critical vulnerabilities in controlled AI releases and supply-chain security. Security researcher Pia Hüsch from the Royal United Services Institute noted that while no company is ever completely secure and humans are often the weakest link, the incident "really illustrates how wide the circle of people who may be able to do this is, even if they don't have super technically sophisticated means"

3

.

Ram Varadarajan, CEO at deception-tech firm Acalvio, observed: "The Mythos breach didn't require a sophisticated attack. It just required a contractor, a URL pattern, and a day-one guess, which means the 'controlled release' model failed at its weakest link before the model's capabilities were ever the issue"

5

.

The fact that the breach was uncovered by a reporter rather than Anthropic raises questions about the company's monitoring capabilities. Security experts note that Anthropic should have the means to log and track model use, which should make it possible to stop unauthorized or malicious access, especially for a highly limited rollout

3

.

Scrutiny Over Mythos Capabilities and Marketing Hype

Early analysis from organizations with authorized access suggests Mythos may not be as revolutionary as Anthropic's messaging implied. Alan Woodward at the University of Surrey explained: "The AI is not necessarily capable of finding vulnerabilities that a human wouldn't, but it's just so much faster, thorough and relentless. Hence it's finding vulnerabilities that humans have missed"

2

.

The UK's AI Security Institute tested Mythos and found it capable of attacking only "small, weakly defended and vulnerable enterprise systems" with no indication that truly secure software or networks would be at significant risk, though the institute warned these capabilities are improving rapidly

2

.

Security expert Davi Ottenheimer characterized the situation as "a legitimate technological capability, reframed as civilisational threat, by a party that benefits from the reframing"

2

. Tim Mackey from Black Duck noted that "Anthropic's marketing message for Mythos was effectively a challenge, not dissimilar to a capture-the-flag exercise, where success includes claims of unauthorized access to Mythos"

5

.

The incident underscores broader challenges in balancing AI development transparency with security concerns, particularly as organizations navigate vulnerability research in an era where AI models accelerate both offensive and defensive capabilities.

Today's Top Stories

TheOutpost.ai

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

Instagram logo
LinkedIn logo
Youtube logo
© 2026 TheOutpost.AI All rights reserved