12 Sources
12 Sources
[1]
In the Wake of Anthropic's Mythos, OpenAI Has a New Cybersecurity Model -- and Strategy
OpenAI on Tuesday announced the next phase of its cybersecurity strategy and a new model specifically designed for use by digital defenders, GPT-5.4-Cyber. The news comes in the wake of an announcement last week by competitor Anthropic that its new Claude Mythos Preview model is only being privately released for now -- because, the company says, it could be exploited by hackers and bad actors. Anthropic also announced an industry coalition, including competitors like Google, focused on how advances in generative AI across the field will impact cybersecurity. OpenAI seemed to be seeking to differentiate its message on Tuesday by striking a less catastrophic tone and touting its existing guardrails and defenses while hinting at the need for more advanced protections in the long term. "We believe the class of safeguards in use today sufficiently reduce cyber risk enough to support broad deployment of current models," the company wrote in a blog post. "We expect versions of these safeguards to be sufficient for upcoming more powerful models, while models explicitly trained and made more permissive for cybersecurity work require more restrictive deployments and appropriate controls. Over the long term, to ensure the ongoing sufficiency of AI safety in cybersecurity, we also expect the need for more expansive defenses for future models, whose capabilities will rapidly exceed even the best purpose-built models of today." The company says that it has homed in on three pillars for its cybersecurity approach. The first involves so-called "know your customer" validation systems to allow controlled access to new models that is as broad and "democratized" as possible. "We design mechanisms which avoid arbitrarily deciding who gets access for legitimate use and who doesn't," the company wrote on Tuesday. OpenAI is combining a model where it partners with certain organizations on limited releases with an automated system introduced in February, known as Trusted Access for Cyber or TAC. The second component of the strategy involves "iterative deployment," or a process of "carefully" releasing and then refining new capabilities so the company can get real-world insight and feedback. The blog post particularly highlights "resilience to jailbreaks and other adversarial attacks, and improving defensive capabilities." Finally, the third focus is on investments that the company says support software security and other digital defense as generative AI proliferates. OpenAI says that the initiative fits into its broader security efforts, including an application security AI agent launched last month known as Codex Security, a cybersecurity grants program that began in 2023, a recent donation to the Linux Foundation to support open source security, and the "Preparedness Framework" that is meant to assess and defend against "severe harm from frontier AI capabilities." Anthropic's claims last week that more capable AI models necessitate a cybersecurity reckoning have been controversial among security experts. Some say the concern is overstated and could feed a new wave of anti-hacker sentiment -- consolidating power even more with tech giants. Others, though, emphasize that vulnerabilities and shortcomings in current security defenses are well known and really could be exploited with new speed and intensity by an even broader range of bad actors in the age of agentic AI.
[2]
OpenAI Has a New GPT-5.4-Cyber Model. Here's Why You Can't Use It
OpenAI has a new AI model called GPT 5.4-Cyber, but it's not coming to your ChatGPT. At least not yet. Instead, the company is doing a limited release to verified cybersecurity testers, according to a blog post shared on Tuesday. These experts will put the model through its paces to identify gaps and potential jailbreaks before the model is released to the wider public. OpenAI uses the feedback from these testers for "understanding the differentiated benefits and risks of specific models, improving resilience to jailbreaks and other adversarial attacks, and improving defensive capabilities -- while mitigating harms," the company said. The model release is part of a ramped-up version of OpenAI's Trusted Access for Cyber program, which allows verified cybersecurity professionals and organizations to get early access to models for defense and prevention work. This is a common cybersecurity practice, one made all the more valuable and necessary because of AI. Cyber attackers and defenders alike are armed with AI tools, making cybersecurity an increasingly AI versus AI landscape. When it comes to securing specific AI models, the companies that make them are convinced the latest models are so dangerously powerful that they require extra security. That was the logic behind Anthropic's Project Glasswing, announced last week. Anthropic's next-generation model, Claude Mythos Preview, is apparently so powerful that the company says it has already found security vulnerabilities "in every major operating system and web browser," according to a blog post. Unlike Claude Mythos Preview, which Anthropic said is an entirely new model, OpenAI's GPT-5.4-Cyber is a fine-tuned version of its existing GPT-5.4 large language model. It's been adjusted to focus specifically on cybersecurity and has lower guardrails for security tasks. Essentially, GPT-5.4-Cyber will be less likely to refuse to perform a risky cybersecurity-related task than the normal versions of GPT-5.4. That's so experts can see if and how it could be weaponized by bad actors. The timing of GPT 5.4-Cyber likely isn't coincidental. This is the latest chapter in the ongoing battle for dominance between OpenAI and Anthropic. The companies have been clashing all year to prove their AI models are the most capable, particularly going after government and enterprise contracts. Anthropic kicked off the race with its Claude Cowork and Code tools, which knocked legacy tech companies (and their stock prices) off balance with their agentic abilities. OpenAI was quick to follow with improvements to its Codex coding platform and models, killing off its AI video app Sora to refocus the company's resources.
[3]
OpenAI Releases Cyber Model to Limited Group in Race With Mythos
OpenAI is letting a select group of users access a new artificial intelligence model that's meant to be more adept at spotting software security vulnerabilities, one week after rival Anthropic PBC announced a limited release of an AI tool called Mythos. The ChatGPT maker said Tuesday that it's beginning to roll out GPT-5.4-Cyber, which is aimed at finding issues in software so organizations can fix them. GPT-5.4-Cyber also places fewer constraints on the ways users can probe the model for that task, OpenAI said. The model will be offered to some participants of OpenAI's Trusted Access for Cyber program, which the company rolled out in February to let certain customers and cybersecurity professionals try its most capable offerings. OpenAI plans to increase the number of participants in the early access program. Initially, it will let hundreds of users test out the new model, before expanding that to thousands in the coming weeks. Get the Tech Newsletter bundle. Get the Tech Newsletter bundle. Get the Tech Newsletter bundle. Bloomberg's subscriber-only tech newsletters, and full access to all the articles they feature. Bloomberg's subscriber-only tech newsletters, and full access to all the articles they feature. Bloomberg's subscriber-only tech newsletters, and full access to all the articles they feature. Plus Signed UpPlus Sign UpPlus Sign Up By continuing, I agree to the Privacy Policy and Terms of Service. OpenAI and Anthropic have been racing to develop more advanced AI models that can take on a wide range of capabilities, and to convince businesses to pay for them. But as their models have gotten better at coding, and detecting security issues in code, there have been growing concerns about the technology being misused by criminals and state-backed hackers. Already, AI technology is being used to help enable cyberattacks. Last week, Anthropic introduced an AI model called Mythos that it said specializes in identifying and exploiting vulnerabilities across operating systems and web browsers. The company decided to limit the release to a select group of trusted partners, including Amazon.com Inc., Apple Inc. and Microsoft Corp., with the goal of letting them use Mythos to spot issues and safeguard their systems. Mythos quickly sparked concerns among financial firms and government agencies. During a meeting last week with Wall Street leaders, summoned by US Treasury Secretary Scott Bessent and Federal Reserve Chair Jerome Powell, executives were warned that they should take the Mythos model seriously, Bloomberg News reported. The Treasury Department's technology team is also seeking to gain access to Mythos so it can begin hunting for vulnerabilities.
[4]
OpenAI unveils GPT-5.4-Cyber a week after rival's announcement of AI model
April 14 (Reuters) - OpenAI on Tuesday unveiled GPT-5.4-Cyber, a variant of its latest flagship model fine-tuned specifically for defensive cybersecurity work, following rival Anthropic's announcement of frontier AI model Mythos. Mythos, announced on April 7, is being deployed as part of Anthropic's "Project Glasswing", a controlled initiative under which select organizations are permitted to use the unreleased Claude Mythos Preview model for defensive cybersecurity purposes. It has found "thousands" of major vulnerabilities in operating systems, web browsers and other software. OpenAI, creator of popular chatbot ChatGPT, said that GPT-5.4-Cyber will initially be rolled out on a limited basis to vetted security vendors, organizations and researchers because of its more permissive design. The company is also expanding its Trusted Access for Cyber program to thousands of verified individual defenders and hundreds of teams protecting critical software, it said in a post on its website. OpenAI is adding new tiers to its TAC program, which was launched in February, with higher levels of verification unlocking more powerful capabilities. Users approved for the highest tier will gain access to GPT-5.4-Cyber, which has fewer restrictions on sensitive cybersecurity tasks such as vulnerability research and analysis. Reporting by Juby Babu in Mexico City; Editing by Maju Samuel Our Standards: The Thomson Reuters Trust Principles., opens new tab
[5]
OpenAI releases new cyber security model to limited group of customers
OpenAI has released a model focused on cyber security to a select group of customers, a week after a similar move by rival Anthropic, amid rising concern about AI's ability to exploit software vulnerabilities. The model, called GPT-5.4-Cyber, is designed to autonomously find flaws or bugs in software, alerting cyber security professionals to fix the issues before they are exploited by bad actors. Its launch on Tuesday comes a week after rival Anthropic released its Mythos model, which has similar capabilities, to a small number of organisations. AI labs are concerned that these powerful new models could help hackers and render many current cyber security defences obsolete. The launch of Anthropic's model also knocked software stocks as investors were spooked by how such tools would affect the sector. "Cyber risk is already here and accelerating, but we can act," OpenAI said in a blog announcing its model. "Digital infrastructure has already been vulnerable for years, before advanced AI even came along." Anthropic has said Mythos has already detected thousands of severe vulnerabilities, including in "every major operating system and web browser", some of which had been undetected for decades. An earlier OpenAI cyber security product called Codex Security, released in March, had contributed to fixing more than 3,000 critical and high-priority vulnerabilities, the company said. GPT-5.4-Cyber has been trained to have fewer restrictions than generally available OpenAI models in order to increase its abilities. Both Anthropic and OpenAI have said these tools can help as a line of defence against hackers, who are using AI to find and exploit cyber weaknesses. The AI cyber security arms race has caused concern among governments and financial institutions. Last week, US Treasury secretary Scott Bessent and Federal Reserve chair Jay Powell summoned some of the largest US banks to discuss the cyber risks Mythos posed. Financial regulators and banks have also discussed the model. OpenAI on Tuesday said its new model would be released to members of a trusted access programme for cyber security that it created in February. Customers and professionals are vetted and go through a security process before being granted access to the programme. "Our goal is to make these tools as widely available as possible while preventing misuse," it added. OpenAI said the new model was available only to the highest tier of the programme, but it would seek to expand access in future. The company has not yet disclosed any of its partners.
[6]
OpenAI's new GPT-5.4-Cyber can reverse engineer binaries, and it wants thousands of defenders using it
* OpenAI expands Trusted Access for Cyber to thousands of verified defenders and hundreds of critical-defense teams. * GPT-5.4-Cyber lowers refusal limits and adds binary reverse-engineering for malware and vulnerability analysis. * Trusted Access for Cyber sign-ups open; apply to join TAC for GPT-5.4-Cyber and related defensive tools. AI has had a very strange effect on the world of cybersecurity. Using artificial intelligence to spot, stop, and prevent virus infections is very useful, and they're especially good at identifying if an email is from a legitimate business or a scammer. At the same time, malicious actors are using AI to craft better viruses than ever before, and are especially good at creating realistic voices and messages to fool people. Fortunately, the 'blue team' in cybersecurity got a huge boost earlier this week, after the AI giant Anthropic revealed Mythos for defense purposes and Glasswing to bring a framework to fruition. Now, ChatGPT has revealed its own takes on Anthropic's releases, with GPT‑5.4‑Cyber as the model and an expanded Trusted Access for Cyber (TAC) as the framework. I cancelled my ChatGPT, Perplexity, and Gemini subscriptions for Claude -- and I should have sooner Wish I did this sooner. Posts 50 By Mahnoor Faisal OpenAI reveals its cybersecurity plans soon after Anthropic's own The company is getting serious Over on the OpenAI blog, the company has laid out its plans for its work in the cybersecurity world. It wants to expand its TAC framework to "thousands of verified individual defenders and hundreds of teams responsible for defending critical software." That way, the people who are most in need of AI-powered cybersecurity tools will now have an easier time accessing them. The end goal is to make these specialist tools "as widely available as possible" without handing over the front door keys to malicious actors to exploit. There's also a new model called GPT-5.4-Cyber. While it's always exciting when OpenAI announces a new model, you won't be using this one to book flights or generate images. As the blog post describes it: This is a version of GPT‑5.4 which lowers the refusal boundary for legitimate cybersecurity work and enables new capabilities for advanced defensive workflows, including binary reverse engineering capabilities that enable security professionals to analyze compiled software for malware potential, vulnerabilities and security robustness without needing access to its source code. If you're a part of TAC, or you want to become a part of it, you can sign up and read more about these changes over on the Trusted Access for Cyber website. Anthropic just dropped its core AI safety promise, and that should worry you History doesn't repeat itself, but AI companies sure do. Posts 1 By Mahnoor Faisal
[7]
Like Anthropic, OpenAI Will Share Latest Technology Only With Trusted Companies
The maker of ChatGPT announced the limited release of GPT-5.4-Cyber, a technology designed to find security holes in software. A week after Anthropic said it would limit the release its latest artificial intelligence technology to a small number of trusted organizations because of cybersecurity concerns, OpenAI said on Tuesday that it, too, was sharing a similar technology only with a group of partners. OpenAI, the maker of ChatGPT, said in a blog post that it would initially share a new A.I. model called GPT‑5.4‑Cyber with hundreds of organizations, before expanding the release to thousands of additional partners in the coming weeks. "Our goal is to make these tools as widely available as possible while preventing misuse," the company said. "We aim to make advanced defensive capabilities available to legitimate actors large and small, including those responsible for protecting critical infrastructure, public services, and the digital systems people depend on every day." Like Anthropic's technology, Claude Mythos Preview, GPT-5.4-Cyber is designed to identify security holes in software. Like other tools developed across the long history of cybersecurity, the technology can be used to both attack computer networks and defend them. By releasing the technology to a smaller group, OpenAI, like Anthropic, hopes to give defenders an edge over attackers. Before Anthropic unveiled Mythos last week, Zico Kolter, an OpenAI board member, called for such an approach in an interview with The New York Times. "Four or five months ago, we had a step change in what these systems could do," said Dr. Kolter, a professor of computer science at Carnegie Mellon University who specializes in security and A.I. But security experts disagree on the best way to handle such technologies. If they are not widely distributed from the beginning, some argue, they will ultimately pose a greater security risk because fewer organizations will be able to defend themselves using the most powerful systems. Over the past several months, products from the leading A.I. companies have grown more effective in areas like math and computer programming. Because they are adept at coding, they have a knack for finding security vulnerabilities in widely used software. Companies like OpenAI have also honed technologies specifically for this task. (The Times sued OpenAI and Microsoft in 2023 for copyright infringement of news content related to A.I. systems. The two companies have denied those claims.) OpenAI said it would share its new systems with hundreds of members of its Trusted Access for Cyber program, which it unveiled in February as a way to share technologies with cybersecurity professionals and other partners. The company also said it would reduce cybersecurity-related guardrails on its systems so that professionals could more easily use them to find security vulnerabilities. But as it shares its technologies, it will also work to verify the identity of users in an effort to prevent misuse. Last week, Anthropic limited the release of Claude Mythos to about 40 companies and organizations that maintain critical infrastructure, including the tech giants Apple, Amazon, Microsoft and Google, as well as the Linux Foundation, which oversees the Linux operating system, freely available software that is widely used across the internet.
[8]
OpenAI unveils GPT‑5.4‑Cyber, an AI model for defensive cybersecurity - 9to5Mac
OpenAI has announced a new AI model called GPT-5.4-Cyber. Similar to Anthropic's Claude Mythos, this new "cyber-permissive" variant of its GPT-5.4 is built for defensive cybersecurity and not public use. OpenAI says that its new GPT-5.4-Cyber variant of GPT-5.4 is specifically meant to prepare the way for more capable models coming this year. In preparation for increasingly more capable models from OpenAI over the next few months, we are fine-tuning our models specifically to enable defensive cybersecurity use cases, starting today with a variant of GPT‑5.4 trained to be cyber-permissive: GPT‑5.4‑Cyber. Access to GPT-5.4-Cyber is limited to "the highest tier" of "users willing to work with OpenAI to authenticate themselves as cybersecurity defenders." OpenAI says this is because GPT-5.4-Cyber is "purposely fine-tuned for additional cyber capabilities and with fewer capability restrictions." This is a version of GPT‑5.4 which lowers the refusal boundary for legitimate cybersecurity work and enables new capabilities for advanced defensive workflows, including binary reverse engineering capabilities that enable security professionals to analyze compiled software for malware potential, vulnerabilities and security robustness without needing access to its source code. Because this model is more permissive, we are starting with a limited, iterative deployment to vetted security vendors, organizations, and researchers. The rollout is part of an expanded version of Trusted Access for Cyber, a cybersecurity initiative launched by OpenAI earlier this year. The company highlights two methods for gaining access to Trusted Access for Cyber:
[9]
OpenAI expands access to cyber AI as hacking risks grow
Why it matters: The roadmap coincides with the release of a new model variant, GPT-5.4-Cyber, designed to assist with defensive cybersecurity tasks and be more permissive for vetted users. * Axios first reported on the new cybersecurity product. Between the lines: OpenAI is shifting its approach to cyber risk to focus less on restricting what models can do and more on verifying who gets access to the most sensitive capabilities. * The company says it aims to make tools "as widely available as possible while preventing misuse" through identity verification and monitoring systems, according to a blog post. * OpenAI plans to expand access to thousands of individuals and hundreds of security teams through its already-established Trusted Access for Cyber program, provided they complete verification checks. The intrigue: OpenAI's approach differs from Anthropic's more restrictive rollout, where only about 40 organizations are getting access to Mythos Preview. * Anthropic warned that its model was so adept at its finding and exploiting security flaws that it was simply too dangerous to release widely. * OpenAI is responding to similar security risks by trying to make its tools more widely available for defensive cyber work while still preventing nefarious actors from accessing them. It's a difficult balance to strike. Zoom in: OpenAI is adding new tiers to its Trusted Access for Cyber program, which launched earlier this year, with higher levels of verification unlocking more powerful capabilities. * Users approved for the highest tier will gain access to GPT-5.4-Cyber, which has fewer restrictions on sensitive cybersecurity tasks, such as vulnerability research and analysis. * The model is designed to reduce "unnecessary friction" for legitimate security work, after some cyber partners said they ran into issues with earlier GPT models sometimes refusing to answer dual-use cyber queries. What they're saying: "This is a team sport, we need to make sure that every single team is empowered to secure their systems," Fouad Matin, a cyber researcher at OpenAI, told reporters. "No one should be in the business of picking winners and losers when it comes to cybersecurity." Yes, but: The rollout will be gradual. OpenAI says initial access to the more permissive model will be limited to vetted security vendors, organizations and researchers, but broader availability will scale over time. * The company also expects onboarding to take time as it reviews and verifies users. The intrigue: OpenAI is not currently offering GPT-5.4-Cyber access to U.S. government agencies, but the company told reporters it is in ongoing discussions and will evaluate access through internal governance and safety review processes. Reality check: Some security experts argue that many vulnerabilities identified by AI tools are not necessarily novel or easily exploitable. * However, the speed at which these models are finding security flaws -- and their rapid pace of advancement -- is worrying government officials and global business leaders. What to watch: Running models with these capabilities requires a lot of computing power. Not everyone will be willing to pay the price to run them on their environments.
[10]
OpenAI launches GPT-5.4-Cyber model for vetted security professionals - SiliconANGLE
OpenAI launches GPT-5.4-Cyber model for vetted security professionals OpenAI Group PBC today announced the launch of GPT-5.4-Cyber, a fine-tuned variant of its GPT-5.4 model designed for defensive cybersecurity work and also announced a significant expansion of its Trusted Access for Cyber program to thousands of verified security professionals. The new model has been purpose-built to lower refusal boundaries for legitimate cybersecurity tasks, or in the words of OpenAI, is "cyber-permissive" and adds capabilities not available in the standard version of GPT-5.4. Among the new capabilities is binary reverse engineering, a feature that allows security professionals to analyze compiled software for malware, vulnerabilities and security weaknesses without needing access to source code. Due to its permissive nature, OpenAI is limiting initial access to vetted security vendors, organizations and researchers. Access runs through the company's Trusted Access for Cyber program, which OpenAI launched in February alongside a $10 million cybersecurity grant program. The program now adds tiered verification levels, with the highest tier unlocking GPT-5.4-Cyber. Individual users can verify their identity at chatgpt.com/cyber and enterprises can request access through their OpenAI representative. Customers already enrolled in the program can apply for higher tiers separately. OpenAI is pitching the release as preparation for more capable models expected later this year, saying that it is "fine-tuning our models specifically to enable defensive cybersecurity use cases, starting today with a variant of GPT-5.4 trained to be cyber-permissive." The company added that it expects versions of its current safeguards to be sufficient for upcoming, more powerful models, while more permissive, cyber-specific variants will require stricter deployment controls. OpenAI also cited progress from its Codex Security product, which launched in private beta six months ago and as a research preview earlier this year. OpenAI said Codex Security has contributed to fixes for more than 3,000 critical and high-severity vulnerabilities across the ecosystem since its recent broader launch. The release comes one week after Anthropic PBC introduced Mythos, a new AI model that demonstrated strong cybersecurity capabilities, to a limited group of roughly 40 organizations. OpenAI's rollout is broader, with the company targeting thousands of individual defenders and hundreds of security teams. OpenAI also noted in its announcement that capture-the-flag benchmark performance across its models improved from 27% on GPT-5 in August 2025 to 76% on GPT-5.1-Codex-Max in November 2025 and said it is planning and evaluating future releases "as though each new model could reach 'High' levels of cybersecurity capability" under its Preparedness Framework. The announcement comes amid a broader push that reflects a shift in how AI developers are approaching cybersecurity risk, moving away from blanket capability restrictions toward identity-based access controls. OpenAI said its goal is to make advanced defensive tools "as widely available as possible while preventing misuse" through automated verification systems rather than manual gatekeeping decisions. The Trusted Access for Cyber program is also part of a wider ecosystem investment that includes contributions to open-source security initiatives and free security scanning for open-source projects through Codex for Open Source, which OpenAI said has reached more than 1,000 projects to date.
[11]
OpenAI Expands Cybersecurity Program Before Deploying New Models | PYMNTS.com
By completing this form, you agree to receive marketing communications from PYMNTS and to the sharing of your information with our sponsor, if applicable, in accordance with our Privacy Policy and Terms and Conditions. The company is scaling up the Trusted Access for Cyber (TAC) program to thousands of verified individuals and hundreds of teams responsible for defending critical software, it said in a Tuesday (April 14) blog post. "In preparation for increasingly more capable models from OpenAI over the next few months, we are fine-tuning our models specifically to enable defensive cybersecurity use cases, starting today with a variant of GPT-5.4 trained to be cyber-permissive: GPT-5.4-Cyber," OpenAI said in the post. Because AI tools can be used by attackers as well as defenders, OpenAI has safeguards that include making the tools as widely available as possible while preventing misuse, putting the systems into the world carefully and updating them as needed, and supporting defenders across the ecosystem, according to the post. To expand access to TAC, OpenAI is adding new levels of identity verification. While the company launched the program in February with automated identity verification, it is now expanding it by introducing additional tiers of access for people who are willing to work with OpenAI to authenticate themselves as cybersecurity defenders, per the post. Individual users can verify their identity, while enterprises can request trusted access for their team. Customers who are already in TAC can request additional tiers of access if they further authenticate themselves as legitimate cyber defenders. "Over the long term, to ensure the ongoing sufficiency of AI safety in cybersecurity, we also expect the need for more expansive defenses for future models, whose capabilities will rapidly exceed even the best purpose-built models of today," OpenAI said in the post. OpenAI initially launched Trusted Access for Cyber as a pilot program in February alongside the release of GPT-5.3-Codex, saying it aimed to accelerate cyber defense research. AI company Anthropic announced April 7 that it has a program called Project Glasswing that allows select partners to gain early access to the upcoming Mythos model positioned for defensive cybersecurity work. It was reported Friday (Feb. 10) that the White House is encouraging banks to use Mythos to identify vulnerabilities.
[12]
OpenAI unveils GPT-5.4-Cyber a week after rival's announcement of AI model
April 14 (Reuters) - OpenAI on Tuesday unveiled GPT-5.4-Cyber, a variant of its latest flagship model fine-tuned specifically for defensive cybersecurity work, following rival Anthropic's announcement of frontier AI model Mythos. Mythos, announced on April 7, is being deployed as part of Anthropic's "Project Glasswing", a controlled initiative under which select organizations are permitted to use the unreleased Claude Mythos Preview model for defensive cybersecurity purposes. It has found "thousands" of major vulnerabilities in operating systems, web browsers and other software. OpenAI, creator of popular chatbot ChatGPT, said that GPT-5.4-Cyber will initially be rolled out on a limited basis to vetted security vendors, organizations and researchers because of its more permissive design. The company is also expanding its Trusted Access for Cyber program to thousands of verified individual defenders and hundreds of teams protecting critical software, it said in a post on its website. OpenAI is adding new tiers to its TAC program, which was launched in February, with higher levels of verification unlocking more powerful capabilities. Users approved for the highest tier will gain access to GPT-5.4-Cyber, which has fewer restrictions on sensitive cybersecurity tasks such as vulnerability research and analysis. (Reporting by Juby Babu in Mexico City; Editing by Maju Samuel)
Share
Share
Copy Link
OpenAI unveiled GPT-5.4-Cyber, an AI model designed to find software security vulnerabilities, just one week after rival Anthropic released its controversial Mythos model. The new model is being released to a limited group of verified cybersecurity professionals through OpenAI's Trusted Access for Cyber program, reflecting growing concerns about AI-powered hacking and the potential for misuse of increasingly capable models.
OpenAI announced GPT-5.4-Cyber on Tuesday, a specialized AI model for cybersecurity designed to identify software security vulnerabilities before malicious actors can exploit them
1
. The announcement comes exactly one week after competitor Anthropic revealed its Claude Mythos Preview model, which the company claimed has already detected thousands of severe vulnerabilities in "every major operating system and web browser"5
. Unlike a completely new architecture, GPT-5.4-Cyber is a fine-tuned version of OpenAI's existing GPT-5.4 large language model, specifically adjusted for defensive cybersecurity work with lower guardrails for security tasks2
.
Source: Axios
The model will be available exclusively through OpenAI's Trusted Access for Cyber program, which launched in February to provide verified cybersecurity professionals early access to advanced models
3
. Initially, hundreds of users will test the new model, with plans to expand to thousands in the coming weeks3
. OpenAI is introducing new tiers to its TAC program, with higher verification levels unlocking more powerful capabilities4
. Users approved for the highest tier will gain access to GPT-5.4-Cyber, which has fewer restrictions on sensitive tasks such as vulnerability research and analysis4
.
Source: CNET
While Anthropic's Mythos announcement emphasized catastrophic risks and led to the formation of Project Glasswing, a controlled initiative restricting access to select organizations including Amazon, Apple, and Microsoft
3
, OpenAI adopted a more measured approach. The company stated that "the class of safeguards in use today sufficiently reduce cyber risk enough to support broad deployment of current models"1
. However, OpenAI acknowledged that models explicitly trained for cybersecurity work require more restrictive deployments and appropriate controls, and that long-term AI safety in cybersecurity will demand more expansive defenses1
.OpenAI outlined three core pillars for its cybersecurity approach. The first involves "know your customer" validation systems designed to democratize access while preventing arbitrary gatekeeping
1
. The second component focuses on iterative deployment, carefully releasing and refining new capabilities to improve resilience to jailbreaks and other adversarial attacks while enhancing digital defense capabilities1
. The third pillar centers on investments supporting application security as generative AI proliferates1
.The rapid succession of these announcements highlights the intensifying AI cyber security arms race between OpenAI and Anthropic. Both companies have been competing throughout the year to prove their models are most capable, particularly targeting government agencies and enterprise contracts
2
. The concern about AI-powered hacking has reached the highest levels of government. Last week, US Treasury Secretary Scott Bessent and Federal Reserve Chair Jerome Powell summoned Wall Street leaders to warn them about taking the Mythos model seriously3
. The Treasury Department's technology team is seeking access to Mythos to hunt for vulnerabilities, while financial institutions and financial regulators have held discussions about the model's implications5
.
Source: SiliconANGLE
Related Stories
OpenAI positioned GPT-5.4-Cyber within its broader security ecosystem, pointing to Codex Security, an application security AI agent launched in March that has contributed to fixing more than 3,000 critical and high-priority vulnerabilities
5
. The company also referenced its cybersecurity grants program that began in 2023, a recent donation to the Linux Foundation supporting open source security, and its Preparedness Framework designed to assess and defend against severe harm from frontier AI capabilities1
.Anthropic's claims that more capable AI models necessitate a cybersecurity reckoning have sparked controversy among security experts. Some argue the concern is overstated and could fuel anti-hacker sentiment while consolidating power with tech giants
1
. Others emphasize that well-known vulnerabilities and shortcomings in current security defenses could indeed be exploited with new speed and intensity by a broader range of bad actors, including state-sponsored hackers, in the age of agentic AI1
. OpenAI acknowledged this reality, stating that "cyber risk is already here and accelerating" and noting that "digital infrastructure has already been vulnerable for years, before advanced AI even came along"5
. The potential for misuse remains a central concern as both companies navigate the delicate balance between enabling defenders and preventing exploitation by malicious actors, making the limited release approach critical for understanding software bugs and testing for jailbreaks before wider deployment.Summarized by
Navi
11 Dec 2025•Policy and Regulation

19 Dec 2025•Technology

06 Feb 2026•Technology
