Anthropic debuts powerful AI model for cybersecurity as tech giants unite under Project Glasswing

Reviewed byNidhi Govil

21 Sources

Share

Anthropic has unveiled Claude Mythos Preview, a frontier AI model so adept at finding software vulnerabilities that it won't be publicly released. Through Project Glasswing, the company is partnering with Microsoft, Apple, Google, Amazon Web Services, and over 40 organizations to deploy the model for defensive security work, having already identified thousands of zero-day vulnerabilities—some decades old—across every major operating system and web browser.

Anthropic Unveils Claude Mythos Preview Through Restricted Partnership

Anthropic on Tuesday released a preview of its new frontier model, Claude Mythos Preview, marking a significant shift in how AI companies approach cybersecurity threats. The model won't see a general release. Instead, Anthropic is deploying it through Project Glasswing, a cybersecurity initiative bringing together more than 40 partner organizations including Microsoft, Apple, Google, Amazon Web Services, Nvidia, the Linux Foundation, Cisco, Broadcom, CrowdStrike, JPMorgan Chase, and Palo Alto Networks

1

2

. The restricted access stems from security concerns about the model's capabilities, which Anthropic describes as having strong agentic coding and reasoning skills that weren't specifically trained for cybersecurity work

3

.

Source: ZDNet

Source: ZDNet

The decision to withhold public access reflects the dual-use nature of the new AI model. While it can identify and help patch software vulnerabilities, the same capabilities could be weaponized by adversaries to exploit those weaknesses before defenders can respond. "We need to prepare now for a world where these capabilities are broadly available in 6, 12, 24 months," Logan Graham, Anthropic's frontier red team lead, told WIRED. "Many of the assumptions that we've built the modern security paradigms on might break"

2

.

Thousands of Zero-Day Vulnerabilities Discovered Across Critical Systems

Over recent weeks, Claude Mythos Preview identified thousands of high-severity vulnerabilities, including some in every major operating system and web browser

5

. Many of these zero-day vulnerabilities are one to two decades old, having evaded detection by traditional security tools and human researchers

1

. The model's vulnerability discovery capabilities extend beyond simple bug identification—it can autonomously develop exploits and conduct penetration testing without human steering

3

.

The scope of what Claude Mythos Preview uncovered is staggering. Anthropic's researchers documented a 27-year-old vulnerability in OpenBSD that could crash systems simply through connection attempts, and a 16-year-old flaw in the FFmpeg library that had been "hit five million times by automated testing tools without ever catching the problem"

5

. The model even wrote a web browser exploit chaining together four vulnerabilities, creating a complex attack that escaped both renderer and operating system sandboxes

5

.

Source: VentureBeat

Source: VentureBeat

Industry Collaboration Addresses AI-Driven Cyber Threats

Project Glasswing represents an unusual level of industry collaboration among fierce competitors. The initiative provides partners with private access to scan both first-party and open-source software systems for code vulnerabilities

1

. Anthropic is committing up to $100 million in usage credits, plus $4 million in direct donations to the Linux Foundation and Apache Software Foundation to support the effort

3

. Partners will ultimately share their learnings so the broader tech industry can benefit from the defensive security work

1

.

Source: FT

Source: FT

"Google is pleased to see this cross-industry cybersecurity initiative coming together," said Heather Adkins, Google's vice president of security engineering. "We have long believed that AI poses new challenges and opens new opportunities in cyber defense"

2

. CrowdStrike's CTO Elia Zaitsev highlighted the urgency: "The window between a vulnerability being discovered and being exploited by an adversary has collapsed. What once took months now happens in minutes with AI"

4

.

Offensive and Defensive Capabilities Raise National Security Questions

Anthropic CEO Dario Amodei explained that while the model wasn't specifically trained for cybersecurity, "as a side effect of being good at code, it's also good at cyber"

2

. Beyond vulnerability discovery, Claude Mythos Preview demonstrates advanced exploit development, endpoint security assessment, system misconfiguration hunting, and the ability to evaluate software binaries without access to source code

2

.

The company has engaged in ongoing discussions with US government officials about the model's offensive and defensive capabilities, though these conversations occur against the backdrop of Anthropic's legal battle with the Trump administration after the Pentagon labeled the AI lab a supply-chain risk

1

3

. Newton Cheng, Anthropic's cyber lead for its frontier red team, confirmed the company had "briefed senior officials in the US government about Mythos and what it can do"

3

.

Leaked Documents and Security Incidents Shadow Launch

The existence of Claude Mythos Preview first emerged through a data security incident reported last month by Fortune. A draft blog post about the model—then called "Capybara"—was left in an unsecured cache of documents on a publicly inspectable data lake

1

. The leaked document described it as "by far the most powerful AI model we've ever developed," noting it far exceeded performance in software coding, academic reasoning, and cybersecurity

1

. Anthropic attributed the leak to human error

3

.

The timing proved particularly awkward given Anthropic's recent track record. Last month, the company accidentally exposed nearly 2,000 source code files and over half a million lines of code through a mistake in launching version 2.1.88 of its Claude Code software package, then caused thousands of code repositories on GitHub to be taken down while attempting cleanup

1

. When asked about these incidents, Dianne Penn, a head of product management at Anthropic, told The Verge the company is "taking steps in terms of solidifying our processes"

3

.

What Organizations Should Monitor as AI Reshapes Security

The staggered release approach draws on principles of coordinated vulnerability disclosure, giving developers time to patch bugs before public discussion

2

. However, due to the sheer volume of issues being discovered, fewer than 1% of potential bugs uncovered by the model have been fully patched

5

. This creates an urgent timeline for organizations maintaining critical infrastructure.

Looking ahead, the long-term implications extend beyond immediate patching efforts. As Anthropic and other AI companies face pressure to turn a profit, Project Glasswing could evolve into a paid service providing a new revenue stream—if it proves effective enough for companies to maintain usage

3

. More critically, the initiative aims to prepare the software industry for a future where models of this capability class become commonplace, fundamentally altering assumptions underlying modern security practices

5

.

Today's Top Stories

TheOutpost.ai

Your Daily Dose of Curated AI News

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

© 2026 Triveous Technologies Private Limited
Instagram logo
LinkedIn logo