Anthropic's Claude AI discovers 22 Firefox vulnerabilities in just two weeks of testing

Reviewed byNidhi Govil

9 Sources

Share

In a security partnership with Mozilla, Anthropic used Claude Opus 4.6 to identify 22 CVEs in Firefox over two weeks—14 classified as high-severity. The AI model detected more vulnerabilities than any single month in 2025, though it struggled to create working exploits. The findings highlight AI's growing role in cybersecurity while raising questions about future safeguards.

Claude AI Detects High-Severity Bugs at Accelerated Speed

Anthropic has identified 22 Firefox vulnerabilities through a security partnership with Mozilla, marking a significant milestone in AI-assisted bug hunting. Using Claude Opus 4.6 over a two-week period in January 2026, the AI company discovered 14 high-severity flaws, seven moderate-severity issues, and one low-severity bug

1

. These findings resulted in 22 CVEs, with most bug fixes implemented in Firefox 148, released in February 2026

4

.

Source: Digit

Source: Digit

The 14 high-severity bugs represent almost a fifth of the 73 high-severity vulnerabilities Mozilla fixed throughout 2025

2

. In fact, Claude AI discovered more Firefox vulnerabilities in two weeks than were reported in any single month in 2025, demonstrating how AI is making it possible to detect severe security vulnerabilities at highly accelerated speeds

2

.

Anthropic and Mozilla Collaboration Targets Complex Codebase

The Anthropic team focused their efforts on Firefox's JavaScript engine before expanding to other portions of the codebase. According to Mozilla engineers Brian Grinstead and Christian Holler, Anthropic approached the Firefox team several weeks ago with a newly developed AI-assisted vulnerability detection system

3

. Despite mixed results with prior AI-assisted bug detection systems, this approach proved different—within hours, platform engineers began landing fixes

3

.

Source: TechSpot

Source: TechSpot

Mozilla selected Firefox for testing because "it's both a complex codebase and one of the most well-tested and secure open-source projects in the world"

1

. The AI model detected a use-after-free bug in the browser's JavaScript engine after just 20 minutes of exploration, which was then validated by human researchers in a virtualized environment

4

. By the end of the effort, Claude Opus 4.6 had scanned nearly 6,000 C++ files and submitted 112 unique reports

4

.

Finding Security Vulnerabilities Proves Easier Than Creating Exploits

While Claude AI excelled at identifying bugs, it performed comparatively poorly at exploiting them. Anthropic spent approximately $4,000 in API credits attempting to develop proof-of-concept exploits, but Claude Opus 4.6 succeeded in only two cases

1

. These were "crude browser exploits" that would be unlikely to work in real-world scenarios due to existing AI safeguards and security features like sandboxing

2

.

Source: Axios

Source: Axios

One successful exploit targeted CVE-2026-2796, a just-in-time miscompilation in the JavaScript WebAssembly component with a CVSS score of 9.8

4

. However, Anthropic emphasized that this exploit only worked within a testing environment with intentionally removed security features

3

. The company noted that the cost of identifying vulnerabilities is cheaper than creating exploits, and the model is better at finding issues than exploiting them

4

.

Web Browser Security Gets New Cybersecurity Engineering Tools

Beyond the 22 CVEs, the AI-assisted approach uncovered 90 additional low-priority bugs, bringing the total to over 100 Firefox bugs discovered in just two weeks

5

. Many of these consisted of assertion failures that overlapped with issues traditionally found through fuzzing, but the AI also identified distinct classes of logic bugs that fuzzers failed to catch

4

.

Mozilla views these findings as "clear evidence that large-scale, AI-assisted analysis is a powerful new addition to security engineers' toolbox"

4

. The organization plans to incorporate this new method into its broader security and development workflow, expecting Claude AI models and other advanced systems to help uncover additional issues in the future

5

.

Concerns About AI Hallucinations and Future Safeguards

While Mozilla has found value in bug-seeking bots, not all open-source projects share this positive experience. Daniel Stenberg, a lead developer at curl, reported "an explosion in AI slop reports," with fewer than one in 20 bugs reported to the company in 2025 being actually real

2

. "The AI chatbots still easily hallucinate security problems," Stenberg said

2

.

Anthropic's approach differs significantly from other AI-driven efforts by incorporating a task verifier to determine if exploits actually work, providing real-time feedback as the tool explores the codebase

4

. The company recently launched Claude Code Security in limited research preview, which can highlight vulnerabilities and suggest targeted software fixes for human review

2

.

Looking ahead, Anthropic acknowledged a critical concern: "Looking at the rate of progress, it is unlikely that the gap between frontier models' vulnerability discovery and exploitation abilities will last very long"

3

. If future language models break through this exploitation barrier, the company stated it will need to consider additional safeguards or other actions to prevent models from being misused by malicious actors

3

. The Red Team collaboration demonstrates both the promise and potential risks of AI in cybersecurity, as the technology continues to advance at a rapid pace.

Today's Top Stories

TheOutpost.ai

Your Daily Dose of Curated AI News

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

© 2026 Triveous Technologies Private Limited
Instagram logo
LinkedIn logo