GPT-5.5 matches Claude Mythos in cybersecurity capabilities as OpenAI plans controlled release

7 Sources

Share

OpenAI's GPT-5.5 has matched Anthropic's Claude Mythos in cybersecurity testing, achieving 71.4% success on expert-level tasks. The UK's AI Security Institute found GPT-5.5 solved a complex reverse engineering challenge in just 10 minutes that took human experts 12 hours. OpenAI is now preparing a controlled release of GPT-5.5-Cyber to critical cyber defenders through its Trusted Access program.

GPT-5.5 Achieves Parity with Claude Mythos in Cybersecurity Evaluations

OpenAI's recently launched GPT-5.5 has demonstrated cybersecurity capabilities on par with Anthropic's heavily restricted Claude Mythos Preview, according to new findings from the UK's AI Security Institute (AISI). The research reveals that GPT-5.5 achieved a 71.4% pass rate on expert-level cybersecurity tasks, marginally higher than Claude Mythos Preview's 68.6%, though within the margin of error

1

. These results suggest that advanced offensive cyberattack capabilities may be emerging as a natural byproduct of improvements in reasoning, coding, and autonomous task completion rather than breakthroughs specific to individual models.

Source: Gadgets 360

Source: Gadgets 360

Since 2023, the AI Security Institute has evaluated frontier AI models through 95 different Capture the Flag challenges designed to test skills in reverse engineering, web exploitation, cryptography, and penetration testing

1

. GPT-5.5's performance marks a significant leap from GPT-5.4, which achieved only 52.4% on the same expert-tier tasks

4

.

Ten Minutes to Solve What Takes Humans Twelve Hours

In one particularly striking demonstration, GPT-5.5 solved a fiendishly difficult reverse engineering challenge that required building a disassembler to decode a Rust binary, reconstructing a custom virtual machine's instruction set, and recovering a cryptographic password through constraint solving. The model completed this task autonomously in 10 minutes and 22 seconds at a cost of $1.73 in API calls

1

4

. A human security expert using professional tools required approximately 12 hours to complete the same challenge.

GPT-5.5 also became only the second model to successfully complete AISI's most demanding test, "The Last Ones" (TLO), a 32-step simulated corporate network attack requiring reconnaissance, credential theft, lateral movement across multiple Active Directory forests, supply-chain pivoting through a CI/CD pipeline, and exfiltration of protected internal databases

4

. GPT-5.5 succeeded in two of 10 attempts, compared to Claude Mythos Preview's three of 10 successes. No previous model had ever succeeded at this test even once

1

.

Safety Guardrails Bypassed Through Jailbreak Techniques

Despite OpenAI's efforts to implement protective measures, AISI researchers identified a universal jailbreak that completely bypassed GPT-5.5's safety guardrails, eliciting harmful content across all malicious cyber queries tested, including in multi-turn agentic settings

4

. The attack required six hours of expert red-teaming to develop. OpenAI subsequently updated its safeguard stack, though configuration issues prevented AISI from verifying the effectiveness of the final version. This vulnerability highlights the challenge of containing dual-use AI capabilities that can both identify software vulnerabilities for defensive purposes and enable malware analysis and exploit generation.

Source: Decrypt

Source: Decrypt

OpenAI Plans Controlled Release Through Trusted Access Program

Sam Altman announced that OpenAI will begin rolling out GPT-5.5-Cyber to critical cyber defenders within days through the company's Trusted Access for Cyber program

3

5

. The model is described as "purposely fine-tuned for additional cyber capabilities and with fewer capability restrictions" compared to standard versions

1

. Access will be restricted to government entities, critical infrastructure operators, security vendors, cloud platforms, and financial institutions

5

.

Source: The Register

Source: The Register

This approach represents a broader distribution strategy than Anthropic's handling of Claude Mythos, which was limited to approximately 50 organizations and will never be publicly available

5

. The controlled release follows OpenAI's February launch of its Trusted Access for Cyber pilot program, which allows security researchers and enterprises to verify their identities and register interest in studying frontier models for "legitimate defensive work"

1

.

Fear-Based Marketing or Genuine Concern?

The timing of OpenAI's announcement has drawn attention given Sam Altman's recent criticism of what he termed "fear-based marketing" in the AI industry. During an appearance on the Core Memory podcast, Altman took aim at Anthropic's restricted release strategy for Claude Mythos, stating: "It is clearly incredible marketing to say, 'We have built a bomb. We are about to drop it on your head. We will sell you a bomb shelter for $100 million'"

1

3

. Now OpenAI finds itself implementing similar access controls, though Altman frames the approach as working "with the entire ecosystem and the government to figure out trusted access for cyber"

3

.

The AI Security Institute's findings suggest the cybersecurity risk may be less about specific model breakthroughs and more about general improvements in AI capabilities. AISI concluded that GPT-5.5's performance indicates rapid improvement in cyber capabilities is likely "a byproduct of more general improvements in long-horizon autonomy, reasoning, and coding" rather than "a breakthrough specific to one model"

1

. If offensive cyber skill emerges naturally from wider advances in AI, further improvements could arrive in quick succession, making containment increasingly difficult.

What This Means for Cybersecurity's Immediate Future

The UK government's annual Cyber Security Breaches Survey, published alongside the AISI findings, found that 43% of businesses suffered a cyber breach or attack in the past 12 months

4

. In response, the government announced £90 million in new funding to boost cyber resilience and is moving forward with the Cyber Security and Resilience Bill. Officials also published guidance urging organizations to prepare for a potential surge in newly discovered software vulnerabilities as AI accelerates the pace at which security flaws can be found and weaponized

4

.

OpenAI's base GPT-5.5 model received a "High" rating on the company's cybersecurity risk scale, falling below the "Critical" threshold needed for developing zero-day exploits autonomously without human intervention

5

. Whether GPT-5.5-Cyber closes this gap remains uncertain, as OpenAI has not published technical benchmarks comparing its specialized variant to Claude Mythos. Leading security professionals continue to express concerns that the ability to find software flaws at scale is becoming increasingly difficult to contain, regardless of access controls

5

.

Today's Top Stories

TheOutpost.ai

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

Instagram logo
LinkedIn logo
Youtube logo
Ā© 2026 TheOutpost.AI All rights reserved