2 Sources
[1]
The patching treadmill: Why traditional application security is no longer enough
Vulnerability backlogs are overwhelming development teams. For all the time I've spent exercising on treadmills, I've always found them faintly demoralizing. You thump-thump-thump over and over again, but get nowhere. It's a lot of effort. You always work up a bit of a sweat, but ultimately feel unfulfilled. This feeling is reinforced the next day, when you have to do it all over again. In many ways, application security is like that treadmill. Once the coding is done, security teams (or customers) find flaws. Scanning tools also find flaws, often resulting in reports that seem never-ending. Coders are constantly yanked away from new development to re-learn what they wrote, locate bugs, patch them, and release fixes. Also: 77% of IT managers say their AI agents are out of control - 5 ways to rein in yours But then, like on the treadmill, the cycle repeats when new code, new dependencies, and new vulnerabilities appear. Because, of course, they will. This frustrating process is often called the find-and-fix cycle. Security and QA teams use vulnerability scanners and penetration tests. When problems are found, as they will be, developers work from the bug reports, set up triage queues, and sometimes dedicate blocks of time to remediation sprints. Find-and-fix isn't so much a development strategy as it is a reactive response to shipping code. The hope is that security flaws (all flaws, really) can be identified and fixed after release, but before they create serious harm or before your customers show up at your door with pitchforks and torches, demanding reliable code. Some security flaws are found so deep in older code that fixing them isn't practical. Code change after code change has been layered on an already shaky, compromised foundation. Getting to the root cause would require tearing everything apart, which would undoubtedly break even more. Also: I asked 5 data leaders about how they use AI to automate - and end integration nightmares That's where another time-honored but suboptimal practice, defend-and-defer, comes into play. Rather than fix deeply entrenched, vulnerable code, programmers and security teams add protective walls around it. Firewalls, runtime protections, monitoring, compensating controls, segmentation, access restrictions, and emergency mitigations all somewhat reduce exposure while the underlying application weakness remains unresolved. But at least there's some defense in place, right? Right? Here's the thing. Find-and-fix and defend-and-defer practices will never completely go away. No matter how good our best practices get, life will find a way. There will always be unexpected behavior. Given the non-deterministic nature of large language models, that possibility is even more true in the age of AI. Also: Nearly half of cybersecurity pros want to quit - here's why Find-and-fix and defend-and-defer practices are no longer sufficient. Software development moves way too fast, especially as developers use more AI assistance to crank out new versions and new capabilities at machine speed. It used to be the case that software delivered updates and new versions periodically. Big releases came out once a year. Updates, maybe, once a quarter. But now, with CI/CD (continuous integration/continuous deployment), the operative word is "continuous." Every tweak, every sprint, every bug fix, every dependency update, every cloud configuration change, every new API integration, and every AI-assisted coding session can break things and introduce new security problems faster than traditional security teams can review them. Also: These 4 critical AI vulnerabilities are being exploited faster than defenders can respond And that focus doesn't even consider mitigation. When security teams review code, whether AI-assisted or not, they often reveal hundreds or thousands of problems that need fixing. The problems are being found faster than developers can realistically fix. Worse, most fixes take developers away from innovation and new code development, resulting in a painful and productivity-killing context switch. That's why most software has a queue of unresolved problems and vulnerabilities that regularly need to be prioritized, re-prioritized, accepted, deferred, or ignored. According to security platform provider Edgescan, network issues take an average of 54 days to fix. Web apps take almost 75 days to fix. The problem is worse at big companies. According to Edgescan's analysis, 45% of large-company vulnerabilities remain unfixed after a full year. This situation is not good. The software might create issues for users. The vulnerabilities could be exploited by attackers, bots, and criminal groups. Known but unpatched vulnerabilities are so popular that information about them is sold to others wishing to break into systems. Also: The biggest AI threats come from within - 12 ways to defend your organization When it comes to breaches, Verizon's 2025 Data Breach Incident Report determined that 20% of threat actors gained initial access to systems through code vulnerabilities, up 34% on the previous year. The other two primary access methods were credential abuse (22%) and phishing attacks (16%). In other words, patching vulnerabilities might have blocked 20% of all breach attacks, but success is not that simple. Here's another stat that reinforces this problem. Security analytics company VulnCheck reported that, "32.1% of KEVs had exploitation evidence on or before the day the CVE was issued, an increase from 23.6% in 2024." In short, bad guys knew about the vulnerabilities, KEV stands for known exploited vulnerabilities, before vendors knew they needed to be fixed. CVEs (common vulnerabilities and exposures) are the mechanism often used to notify and track the resolution of known vulnerabilities. Basically, the VulnCheck stat reported that almost a third of all vulnerabilities were in bad actors' hands and being actively exploited before the developers who could fix the vulnerabilities even found out about them. Unfortunately, we can't just demand that developers patch code with improved speed or productivity. Beyond the physical limits of human coders, or even the enhanced performance but practical limits of our AI overlords, there are practical concerns. Enterprise systems have dependencies, uptime requirements, change-control boards, regulatory constraints, customer commitments, fragile integrations, and teams that may not own the vulnerable code. Also: Rolling out AI? 5 security tactics your business can't get wrong - and why Smaller systems may depend on components or elements out of their control. For example, I woke up one morning this week to find that five of my legacy websites were no longer functioning. Those sites had been working perfectly. They had been unmodified for at least seven years. The hosting operator changed a version of a critical software system without warning, and some of my custom code stopped functioning. It took me a few days to get back up to speed on what my code did, then track down and fix it. And that was with the help of OpenAI Codex. Then there's the issue of prioritization fatigue. When every vulnerability comes in as critical, it's as if nothing is crucial. Did you ever have a day where you prioritized your to-do list, only to realize that you had 30 top-priority tasks? I see you nodding your head. At that point, it's just overwhelming, and no issue stands out. Also: Will AI make cybersecurity obsolete, or is Silicon Valley confabulating again? Even AI-driven vulnerability scans won't help you deal with the challenge. Super tools, like Anthropic Mythos, or even more accessible tools, such as Claude Security or Codex Security, can't really solve the problem. A dashboard full of findings can create the appearance of control, while the underlying engineering practices continue to produce the same defect categories. It's at this point that IT operators often try the defend-and-defer approach using tools like network or application firewalls, intrusion detection and prevention systems, endpoint detection and response, network segmentation, rate limiting, logging and monitoring, runtime application self-protection, or even virtual patching. These "compensating controls" are sometimes essential, but they can become a permanent substitute for fixing root causes. This practice is dangerous because surrounding weak software with a scaffold of security tooling doesn't solve the underlying problem: weak code. Patching after the fact isn't just insecure, it's really expensive. Yes, it's sometimes necessary (like when, a decade after I wrote a line of code using the standards at the time, a much later OS release broke it). But coding defensively, and making fixes while the original code is being developed, is far less time-consuming and painful than identifying, triaging, patching, validating, deploying, and monitoring fixes way after release. It's hard to pin down exactly when "modern development" practices started, because everyone has a different perspective. But it's fair to say that development lifecycles changed when we went from shipping updates on disk to building cloud-centric services. Then the practice changed again in the past few years when AI-assisted development became a transformative force. The fact is, our approach to software development is different from the time when find-and-fix was the way of the world. Application risk now pervades the whole software lifecycle: design choices, coding practices, dependency selection, secrets handling, identity controls, build pipelines, deployment configurations, and runtime exposure. Also: Why enterprise AI agents could become the ultimate insider threat As I've been discussing for the past year, AI has radically changed release timelines, accelerating schedules, and collapsing timelines. Unfortunately, that increase in speed can widen the gap between code creation and security review. If nothing else, the volume of code produced has increased as the time to create code has collapsed. Testing time, on the other hand, has not flattened. I've been working on a Mac app in Claude Code for about four months. The actual code-writing process takes about 20 minutes each session. But because my code uses on-device AI for sophisticated document parsing, the testing takes hours each session. My coding time has collapsed to a mere rounding error, but the testing time now takes the bulk of my development time. Still, without having AI for the initial code-writing process, I probably wouldn't have time to finish this project, whenever that happens. Also: 7 AI coding techniques I use to ship real, reliable products - fast The key problem is that AI-generated code is not necessarily secure code. AI monitoring company Snyk reported that 56.4% of developers frequently encountered security issues in AI-generated code, while 80% ignored or bypassed organizational AI code-security policies. In this article, we've looked at what happens as software production accelerates, but security remains a downstream problem: the treadmill speeds up. More code means more problems, which are found faster than developers can go back and make fixes. To be clear, we will never be able to abandon find-and-fix or defend-and-defer practices. Stuff happens. We'll always need to employ scanning, patching, monitoring, and runtime defense to some degree. But these practices should be migrated to a second-tier safety net. Do vulnerability backlogs feel like a manageable process or an endless queue in your organization? Let us know in the comments below.
[2]
Standard 90-day vulnerability disclosure policy is likely dead thanks to AI, expert warns that AI can weaponize patches in 30 minutes -- LLM-assisted bug-hunting ushers in a new cyberworld order
If you're not integrating LLMs in your development pipeline for security checks, you've already lost. In case you haven't been in the cybersecurity news lately, here's a quick summary: discoveries and exploits of high-profile software vulnerabilities are becoming faster than ever, in part due to AI-assisted code scanning tools. For example, most every Linux distribution recently found itself on the wrong end of the Copy Fail and Dirty Frag privilege escalation vulnerabilities (gaining administrator access with a local account), for which patches hadn't been made widely available as there wasn't enough time between their disclosure and publication. Himanshu Anand, a security researcher, wrote a lengthy blog post explaining why the industry-standard 90-day disclosure window and associated procedure are effectively dead in this AI-powered world, and his conclusions might lead developers and sysadmins to pick up a stiff drink. On the developer side, he suggests programmers to add LLM to their code push, deployment, and dependency-checking steps as a countermeasure, as attackers are already using LLMs to undercover vunerabilities. The crux of the matter is the fact that although a bot isn't necessarily any smarter than a human at programming or hunting for security vulnerabilities, a LLM that can do so at full mental capacity 24/7 and is brutally effective at pattern recognition (built with pattern recognition, if we must). The vast majority of security exploits are rooted in specific bad programming habits, something a bot excels at noticing quickly and repeatedly. Both aforementioned exploits for the Linux kernel took advantage of insecure zero-copy mechanisms (performing calculations on data in-place instead of copying/calculating/replacing). In both cases, although the issues were communicated to the kernel team in advance, they were made public far before the usual 90-day period -- just over a week, in the case of Dirty Frag. Although nobody said it out loud, the general assumption was that white-hat reveals were done with little to no advance warning because the exploits were already in the wild, so there was nothing to gain and everything to lose by keeping them under wraps. To illustrate this point, Anand presents one of his own bug reports to an unnamed e-shop, wherein he found and reported an unpatched security bug that would let attackers buy expensive items for the princely sum of $0. Much to his surprise, he got a reply stating that 10 (!) other researchers had already reported the issue over six weeks. Conferring with a colleague, they noticed that "LLM-assisted hunters were converging on the same bugs almost simultaneously." This conclusion is further backed up by triage engineer @d0rsky, who notes that once a new vulnerability is found, he immediately sees "a wave of duplicate reports within days." Quite poignantly, Dorsky posits: "if researchers can replicate these findings so quickly, what's stopping black-hats from doing the same before the issue is fixed?" Anand further drives the point home by saying he made an exploit for a published and patched vulnerability in the React framework in just 30 minutes using LLM tools. In his conclusion, Anand doesn't mince words, stating that in this new world where non-ethical hackers can so quickly analyze code using AI, the 90-day window protects nobody, and that the usual monthly patch cycles are equally dead, as "[the] 30 day window between vulnerability and fix assumes attackers are slower than your release train." He urges developers to treat "every critical security issue as P0 and fix it immediately," as they can assume that said vulnerability is already under active exploitation. To wit, "if you are reading CVE descriptions while attackers are reading git log --diff-filter=M, you are already behind." Ironically enough, open-source software enjoys high security standards due to code being publicly available for scrutiny and correction, but LLMs are turning that characteristic into a double-edged sword. Having said that, in the OSS world, a patch can also be created and distributed within hours, something the Mozilla team recently proved by posting 423 security fixes in April alone. As for closed-source software, well, let's just say that tireless bots are equally good at decompiling and network scanning as they are at source code analysis, and it's likely enough that Microsoft, Apple, or Google will have their Copy Fail moments sooner rather than later. Do read the entirety of Anand's post, as it's quite elucidative. Follow Tom's Hardware on Google News, or add us as a preferred source, to get our latest news, analysis, & reviews in your feeds.
Share
Copy Link
Security researchers warn that AI-assisted bug hunting has effectively killed the industry-standard 90-day vulnerability disclosure policy. LLMs can now identify and weaponize vulnerabilities within 30 minutes of patches being published, leaving systems exposed to zero-day attacks faster than traditional security teams can respond. The patching treadmill that once frustrated developers has become a full-blown crisis.
The landscape of application security has shifted dramatically as AI and LLMs accelerate both vulnerability discovery and exploitation. Security researcher Himanshu Anand recently declared the industry-standard 90-day vulnerability disclosure policy effectively dead, warning that attackers using AI-assisted bug hunting can now weaponize patches in just 30 minutes
2
. This development marks a fundamental break from traditional application security practices that relied on slower, human-paced review cycles.
Source: Tom's Hardware
Anand demonstrated this threat by creating an exploit for a published React framework vulnerability in 30 minutes using LLM tools
2
. The speed advantage stems from LLMs operating at full capacity around the clock, excelling at pattern recognition to identify the specific bad programming habits that underpin most security exploits. When Anand reported a zero-dollar purchase bug to an e-commerce platform, he discovered 10 other researchers had already flagged the same issue over six weeks, with LLM-assisted hunters converging on identical bugs almost simultaneously.Development teams already struggle with what's known as the patching treadmill—a frustrating cycle where security teams and scanners identify flaws, developers get pulled from new work to patch them, only to repeat the process as new code and dependencies introduce fresh vulnerabilities
1
. The find-and-fix cycle, coupled with defend-and-defer practices that add protective walls around deeply entrenched vulnerable code rather than fixing root causes, creates mounting vulnerability backlogs that overwhelm teams.
Source: ZDNet
According to security platform provider Edgescan, network issues take an average of 54 days to fix while web applications require almost 75 days
1
. The situation worsens at large companies, where 45% of vulnerabilities remain unfixed after a full year. These delays create windows of opportunity that AI now exploits at unprecedented speed. Triage engineer @d0rsky observed that once a new vulnerability surfaces, "a wave of duplicate reports within days" follows, raising the critical question: if researchers can replicate findings so quickly, what stops black-hat actors from doing the same before fixes deploy2
?The acceleration of AI-assisted coding has intensified pressure on security teams. Continuous integration/continuous deployment (CI/CD) means software updates flow constantly rather than quarterly or annually
1
. Every tweak, sprint, dependency update, cloud configuration change, and AI-assisted coding session can introduce security risks faster than traditional teams can review them. When security teams do review code, they often uncover hundreds or thousands of problems, with issues being found faster than developers can realistically address them.Recent Linux kernel vulnerabilities Copy Fail and Dirty Frag illustrated this new cyberworld order. Both exploited insecure zero-copy mechanisms and were disclosed publicly just over a week after discovery—far short of the usual 90-day period—because the exploits were already in the wild
2
. This reality renders monthly patch cycles equally obsolete, as the 30-day window between vulnerability and fix assumes attackers move slower than release schedules.Related Stories
Anand's stark warning to developers: "If you're not integrating LLMs in your development pipeline for security checks, you've already lost"
2
. He urges teams to treat every critical security issue as priority zero and fix immediately, operating under the assumption that vulnerabilities are already under active exploitation. The advice reflects a harsh new reality: reading CVE descriptions while attackers analyze git logs puts defenders perpetually behind.Open-source software faces a double-edged sword in this environment. While publicly available code traditionally enabled high security standards through community scrutiny, LLMs now analyze that same transparency to identify weaknesses at machine speed. Mozilla recently demonstrated the potential response velocity by posting 423 security fixes in April alone
2
. Closed-source software offers no sanctuary, as tireless bots prove equally adept at decompiling and network scanning as source code analysis.The implications extend beyond technical fixes. Organizations must fundamentally rethink security workflows, accepting that the 90-day window protects nobody when ethical hackers and malicious actors alike leverage AI to converge on vulnerabilities simultaneously. The shift demands immediate action on critical flaws rather than queuing them in vulnerability backlogs that assume human-paced exploitation timelines no longer apply.
Summarized by
Navi
22 Apr 2026•Technology

23 Dec 2025•Technology

27 Mar 2026•Technology

1
Technology

2
Policy and Regulation

3
Policy and Regulation
