OpenAI Codex Security scans 1.2M commits, finds 792 critical software vulnerabilities

4 Sources

Share

OpenAI rolled out Codex Security, an AI-powered security agent that scans code repositories to identify and fix vulnerabilities. During beta testing, it analyzed 1.2 million commits and discovered 792 critical findings and 10,561 high-severity issues across major open-source projects. The tool is now available in research preview to Enterprise and Business customers with free usage for the first month.

OpenAI Launches AI-Powered Security Agent to Identify Software Vulnerabilities

OpenAI began rolling out Codex Security on Friday, marking a significant entry into the competitive market for AI-powered application security tools

2

. The AI-powered security agent is designed to find, validate, and propose fixes for software vulnerabilities across code repositories

1

. Available in research preview to ChatGPT Enterprise, Business, and Edu customers via Codex web, the platform offers free usage for the first month as OpenAI tests its capabilities with a broader user base

3

.

Codex Security represents an evolution of Aardvark, which OpenAI unveiled in private beta in October 2025 as a way for developers and security teams to detect and fix security vulnerabilities at scale

1

. Ian Brelinsky, a member of OpenAI's Codex Security team, emphasized the defensive focus: "We wanted to make sure that we're empowering defenders"

2

.

Scanning 1.2 Million Commits Reveals Critical Security Gaps

During beta testing over the last 30 days, Codex Security scanned more than 1.2 million commits across external repositories, identifying 792 critical findings and 10,561 high-severity findings

1

. These discoveries include vulnerabilities in various open-source projects like OpenSSH, GnuTLS, GOGS, Thorium, libssh, PHP, and Chromium

1

. The platform analyzes code repositories, pressure-tests suspected vulnerabilities in sandboxed environments, generates proof-of-concept exploits to confirm impact, and proposes fixes

2

.

OpenAI is supporting open-source security by scanning major repositories, sharing high-confidence findings with maintainers, and reporting vulnerabilities in critical projects

3

. The company plans to expand its support to more open-source maintainers, offering tools and resources to enhance security measures across the ecosystem

3

.

How Frontier Models Automate Code Security Reviews

The AI agent leverages the reasoning capabilities of OpenAI's frontier models and combines them with automated validation to minimize the risk of false positives and deliver actionable fixes

1

. Using these frontier models, Codex Security builds detailed threat models for projects, prioritizing real-world risks and allowing security teams to focus on critical issues

3

.

The agent works in three steps: it analyzes a repository to understand the project's security-relevant structure and generates an editable threat model that captures what it does and where it's most exposed

1

. Once the system context is built, Codex Security uses it as a foundation to identify vulnerabilities and classifies findings based on their real-world impact. The flagged issues are pressure-tested in a sandboxed environment to validate them

1

. The final stage involves proposing fixes that best align with system behavior to reduce regressions and make them easier to review and deploy

1

.

Source: Hacker News

Source: Hacker News

Reducing False Positives Through Autonomous Cybersecurity Solutions

OpenAI's scans on the same repositories over time have demonstrated increasing precision and declining false positive rates, with the latter falling by more than 50% across all repositories

1

. Originally released as a private beta named Aardvark, it identified severe flaws like cross-tenant authentication vulnerabilities and improved accuracy during testing, cutting noise by 84% and reducing false positives by over 50%

3

.

OpenAI stated that Codex Security is designed to improve signal-to-noise by grounding vulnerability discovery in system context and validating findings before surfacing them to users

1

. "By combining agentic reasoning from our frontier models with automated validation, it delivers high-confidence findings and actionable fixes so teams can focus on the vulnerabilities that matter and ship secure code faster," OpenAI said .

Competition Intensifies in AI-First Threat Prevention Market

OpenAI is entering a growing market for AI-enabled code security tools, escalating competition among both traditional application security vendors and rival AI labs

2

. Anthropic made a similar move last month when it introduced Claude Code Security, rattling share prices for traditional cybersecurity vendors

2

. As attackers weaponize AI models, frontier AI labs are increasingly rolling out new ways to help defenders beef up their own security

2

.

Many security executives argue enterprises will likely continue to rely on a mix of vendors, rather than depend solely on the same AI platform provider to both build and secure their systems

2

. However, a new category of AI-first threat prevention platforms is emerging that don't wait for alerts but seek out weak points in code, configurations or behavior and take defensive action automatically .

Why This Matters for Cyber Threats and Remediation

The World Economic Forum found in January that AI is expected to be the most consequential factor shaping cybersecurity strategies this year, with 94% of surveyed executives citing the technology as a force multiplier for defense and offense .

Source: Axios

Source: Axios

Enterprises and investors are shifting toward autonomous remediation as human-speed remediation is no longer sufficient when AI-driven attackers operate in continuous loops .

Brelinsky indicated that code security is just one part of the broader cybersecurity ecosystem, and the company is eying ways to bring more agentic capabilities to defenders

2

. In scenarios where defensive agents remove the need for human intervention on specific vulnerability classes, or compress triage and coordination so engineers focus on higher-order judgment, the common thread is speed .

Today's Top Stories

TheOutpost.ai

Your Daily Dose of Curated AI News

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

© 2026 Triveous Technologies Private Limited
Instagram logo
LinkedIn logo