18 Sources
18 Sources
[1]
Leaked Anthropic Model Presents 'Unprecedented Cybersecurity Risks,' Much to Pentagon's Pleasure
There are two groups that seem absolutely thrilled by the recent revelation that Anthropic is working on a top-secret AI model: the Department of Defense and Anthropic's fundraising team. On Thursday, Fortune reported that Anthropic left a slew of not-yet-public information accessible on its website, leading to the reveal of a supposedly super-powerful, yet-to-be-released model called Claude Mythos. The modelâ€"which is supposedly “by far the most powerful AI model we’ve ever developed," per the text of a leaked, unpublished blog post from Anthropicâ€"blows away benchmarks set by the company's current public model, Claude Opus 4.6. The company confirmed the model's existence to Fortune and called it "a step change and the most capable model we've built to date." The company also basically said it's holding back the model because it might actually be too dangerous to release to the public at this point. Per Fortune, Anthropic said the model is “currently far ahead of any other AI model in cyber capabilities," and presents unprecedented cybersecurity risks that the company seems to want to get out in front of. That revelation has the Pentagon doing victory laps, even though it never gave any indication of having concerns about this type of thing before. The Department of Defense is still carrying a grudge against Anthropic after the company refused to agree to allow its model to be used for domestic surveillance or fully autonomous military weapons. And while it seems the Pentagon is losing its legal caseâ€"a ruling handed down from a judge Thursday temporarily blocked the DoD from labeling Anthropic a security riskâ€"it's still looking to win in the court of public opinion. Following the reveal of Claude Mythos, Under Secretary of War and chief Anthropic antagonist Emil Michael posted, "Ummâ€|hello? Is it not clear yet that we have a problem here?" Michael has been hammering Anthropic for weeks now, calling the company's CEO Dario Amodei a "liar" with a "god complex" who wants to "personally control the US military," and now he's treating the leak as proof that Anthropic can't be trusted. Michael's position is almost certainly not coming from a place of genuine concern. Anthropic was given the green light to handle classified material from the federal government, and most of the military applications for the model are operating within secure environments (or they are supposed to be). The fight has been about the Pentagon wanting to use Anthropic's models even more than the company wants, which doesn't seem like the thing you'd fight for if you thought the models presented security risks. Michael has significant financial ties to other AI firms that are competitors of Anthropic, so it's hard to take his view of the matter at face value. Also, this is the same Pentagon that added a journalist to a Signal group chat where war plans were being shared, so let's not pretend like their security protocols are airtight. Regardless, if we take the leak at face value, it seems like Anthropic is sitting on a big release, though who knows when or if we'll ever see it. It's reportedly very expensive to train and operateâ€"so expensive, in fact, that the company seemingly isn't sure how to deploy it at this point. There's little reason to believe that Anthropic left this information out in the open with the intention of someone finding it. Frankly, it's a pretty embarrassing oversight to leave your content management system accessible without requiring a password, which is how this information all became public. But it's also hard to ignore the fact that this whole situation plays right into the classic AI company playbook of talking up the dangers of a model to highlight how powerful and capable it is. So, it's fitting that around the same time as the leak about the extremely powerful model that blows everyone else away, The Information reported that Anthropic was eyeing going public later this year. That secret, super-powerful model does seem like the kind of thing Anthropic would like potential investors to know about ahead of an IPO. Read More:
[2]
Everyone's worried that AI's newest models are a hacker's dream weapon
The one to watch: Anthropic is privately warning top government officials that its not-yet-released model -- currently branded "Mythos" -- makes large-scale cyberattacks much more likely in 2026. The model allows agents to work on their own with wild sophistication and precision to penetrate corporate, government and municipal systems. It's a hacker's dream weapon. * Jim revealed in his new weekly newsletter for CEOs that one source briefed on the coming models says a large-scale attack could hit this year. Businesses are ripe targets. (C-suite only: Request beta of Jim's newsletter.) Fortune got its hands on an unpublished Anthropic blog post describing Mythos. The post said the model is "currently far ahead of any other AI model in cyber capabilities." * It adds that Mythos "presages an upcoming wave of models that can exploit vulnerabilities in ways that far outpace the efforts of defenders." The threat is no longer theoretical, and will be exacerbated by employees testing agents without realizing they're making it easier for cybercriminals to hack their company. Flashback: Late last year, Anthropic disclosed the first documented case of a cyberattack largely executed by AI -- a Chinese state-sponsored group that used AI agents to autonomously hack roughly 30 global targets, with the AI handling 80-90% of tactical operations independently. * This was before agents got exponentially better and those experimenting with them started to open risky new side doors. Here's why this is different: The new models are even better at powering agents to think, act, reason and improvise on their own without rest or pause or limitation. * Think of a warehouse full of the most sophisticated criminals who never sleep, learn on the fly and persist until successful -- except the warehouse is infinite. * Bad actors can now scale simply with more compute. They aren't limited by finite personnel. A single person can run campaigns that once required entire teams. At the same time, systems are more vulnerable because so many employees are firing up Claude, Copilot or other agentic models -- often at home -- and creating agents of their own. * Oftentimes, they connect to their internal work systems unwittingly, opening a new door for cybercriminals to enter. * The industry has a name for this: "shadow AI." A Dark Reading poll found that 48% of cybersecurity professionals now rank agentic AI as the #1 attack vector for 2026 -- above deepfakes, above everything else. The bottom line: Everyone working at every company in America needs to know right now the dangers of using agents, especially unsupervised, anywhere near sensitive information. Leaders need to hammer this home. * My tech team says this is the biggest threat to Axios right now. * Your workplace can build a safe "playpen" for work-related AI experiments using agents. We're scrambling to finish ours. Go deeper: How Anthropic's Pentagon deal could get revived.
[3]
Meet Claude Mythos: Leaked Anthropic post reveals the powerful upcoming model
The model, now known as "Claude Mythos," was originally uncovered in a report from Fortune. Anthropic has since confirmed the details about the leak to the outlet. The data leak included details about the upcoming release of the Claude Mythos AI model, an "exclusive CEO event," and other internal assets such as PDFs and images, according to Fortune. In total, the leak included nearly 3,000 assets that Anthropic had not previously published. According to Anthropic, the leak was the result of an issue with the content management system (CMS) that the company uses. The data was uploaded to the CMS by the company. However, Anthropic failed to mark the items as private so the data was stored in a publicly accessible data lake. Some of the data was unused assets from past announcements from Anthropic. Other leaked data was internal corporate information such as employee details or info about an invite-only CEO event. Also found within the leaked data, however, was information about Claude Mythos in an unpublished blog post draft. According to the leaked draft about Claude Mythos, Anthropic has completed its training of the model and refers to it as "by far the most powerful AI model we've ever developed." In a statement provided to Fortune after the leak, an Anthropic spokesperson described Claude Mythos as an AI performance "step change" and said the model is "the most capable we've built to date." Anthropic says that Claude Mythos is currently in the trial stage and available to select "early access customers." In addition to the new AI model, the leaked post also mentions a new, unreleased AI model tier from Anthropic called Capybara. Anthropic currently offers three different tier levels for each of its AI models: Opus, Sonnet, and Haiku. Opus is currently its most powerful and most expensive tier whereas Haiku is the smallest and least expensive. Capybara would sit even above Opus as Anthropic's largest and most intelligent AI model tier. The leaked post also details Anthropic's concerns about the cybersecurity risks that Claude Mythos brings to the table, too. It appears that Anthropic believes that Claude Mythos could be used by hackers to run cyberattacks. "In preparing to release Claude Capybara, we want to act with extra caution and understand the risks it poses -- even beyond what we learn in our own testing," reads the leaked Anthropic post. "In particular, we want to understand the model's potential near-term risks in the realm of cybersecurity -- and share the results to help cyber defenders prepare." Anthropic described its unreleased model as "currently far ahead of any other AI model in cyber capabilities" and that the model "presages an upcoming wave of models that can exploit vulnerabilities in ways that far outpace the efforts of defenders." The company said in the leaked draft blog that it was providing early access to organization to give them "a head start in improving the robustness of their codebases against the impending wave of AI-driven exploits."
[4]
Anthropic Just Leaked Upcoming Model With "Unprecedented Cybersecurity Risks" in the Most Ironic Way Possible
Can't-miss innovations from the bleeding edge of science and tech As companies continue to burn through billions of dollars by running massively resource-hungry AI models -- and only passing on a fraction of the costs to consumers and enterprise clients -- the AI race shows no signs of slowing down. On Thursday, a data leak caused by a major security lapse in its public-facing content management system revealed that Anthropic is working on a powerful new model release. The company has since officially acknowledged the new project, dubbed "Claude Mythos," with a spokesperson describing it to Fortune as a "step change" in AI proficiencies and the "most capable we've built to date." The spokesperson said it's a "general purpose model with meaningful advances in reasoning, coding, and cybersecurity." In an enormously ironic twist, a draft blog obtained by Fortune, which was "available in an unsecured and publicly-searchable data store," claimed that the new model "poses unprecedented cybersecurity risks." In other words, let's hope the new model wasn't responsible for the security of Anthropic's company blog. It's a major test for the company, which has received significant media attention as of late for its Claude Code and Claude Cowork tools, the successes of which appear to have rattled Anthropic's competitors, including OpenAI, to their core. The leaks also revealed a "new tier" of AI models, dubbed Capybara. Mythos appears to be part of this new tier, but how Capybara fits in with Anthropic's existing tiers -- Opus, Sonnet, and Haiku, in decreasing size, capability, and cost -- remains to be seen. "Compared to our previous best model, Claude Opus 4.6, Capybara gets dramatically higher scores on tests of software coding, academic reasoning, and cybersecurity, among others," the leaked blog reads, as quoted by Fortune. While it may score higher in cybersecurity tests, it could simultaneously represent a major challenge for existing cybersecurity defenses, the company warned. "In preparing to release Claude Capybara, we want to act with extra caution and understand the risks it poses -- even beyond what we learn in our own testing," the company wrote in the leaked blog post. "In particular, we want to understand the model's potential near-term risks in the realm of cybersecurity -- and share the results to help cyber defenders prepare." The model "presages an upcoming wave of models that can exploit vulnerabilities in ways that far outpace the efforts of defenders," Anthropic boasted. The risks appear to have been real enough for cybersecurity stocks to plunge on Friday, following the latest news. Anthropic has also previously admitted that hackers used its Claude AI model to automate cybercrimes targeting banks and governments. According to the company's November blog post, a Chinese state-sponsored group exploited the AI's agentic capabilities to infiltrate "roughly thirty global targets and succeeded in a small number of cases" by "pretending to work for legitimate security-testing organizations" to sidestep Anthropic's AI guardrails. Reality check: a frontier AI company is working on what it claims to be the next big thing that's more capable than anything that's come before it is pretty standard fare, and it remains to be seen whether Claude Mythos will actually represent a major "step change" in practice, outside of a carefully curated testing environment. Case in point, OpenAI's long-awaited GPT-5 model turned out to be a major letdown when it was released in August, falling well short of the company's lofty promises.
[5]
Anthropic confirms powerful new AI model after data leak
Anthropic is testing a new AI model called "Claude Mythos," the company confirmed after Fortune reported that draft materials about the in-development system had been left in an unprotected, publicly accessible data store on the company's website. The company's spokesperson characterized the new system as "a step change" in AI performance, adding that it was "the most capable we've built to date." The company said it was moving carefully with the rollout because of the model's power, and that a small group of customers currently has early access. Fortune reporter Bea Nolan identified the exposed data. Two cybersecurity researchers -- Roy Paz of LayerX Security and Alexandre Pauwels of the University of Cambridge -- separately assessed the documents at Fortune's request. Pauwels counted nearly 3,000 files tied to Anthropic's blog that had not previously appeared on any of the company's public-facing pages. Anthropic attributed the exposure to a configuration error in one of its external content management tools, calling it "human error." After Fortune contacted the company, Anthropic restricted access to the data store. The draft blog post applied both names -- "Claude Mythos" and "Capybara" -- to what the document indicated was a single model. According to the document, the system outperformed Claude Opus 4.6 across several benchmarks -- including cybersecurity, software coding, and academic reasoning -- and would occupy a new tier above Opus in Anthropic's model lineup. The document also flagged the model's high running costs and said a public launch had not yet been scheduled. The documents outlined significant cybersecurity risks associated with the model. The draft blog characterized the system as more advanced in cybersecurity tasks than any competing AI model and warned that it could allow attacks to scale faster than defenders could counter them. Anthropic said the early-access rollout would focus on cyber defense organizations, giving them time to reinforce their systems ahead of a broader release. The exposed files also contained details of a planned private summit for European business leaders, to be held at a U.K. country manor. Anthropic CEO Dario Amodei is set to attend. Anthropic confirmed the event in a statement, saying it was one in a series of gatherings the company has held for business leaders over the past year.
[6]
Exclusive: Anthropic is testing 'Mythos' its 'most powerful AI model ever developed' | Fortune
AI company Anthropic is developing and has begun testing with early access customers a new AI model more capable than any it has released previously, the company said, following a data leak that revealed the model's existence. An Anthropic spokesperson said the new model represented "a step change" in AI performance and was "the most capable we've built to date." The company said the model is currently being trialed by "early access customers." Descriptions of the model were inadvertently stored in a publicly-accessible data cache and were reviewed by Fortune. A draft blog post that was available in an unsecured and publicly-searchable data store prior to Thursday evening said the new model is called "Claude Mythos" and that the company believes it poses unprecedented cybersecurity risks. The same cache of unsecured, publicly discoverable documents revealed details of a planned, invite-only CEO summit in Europe that is part of the company's drive to sell its AI models to large corporate customers. The AI lab left the material, including what appeared to be a draft blog post announcing a new model, in an unsecured, public data lake, according to documents separately located and reviewed by Roy Paz, a senior AI security researcher at LayerX Security, a computer and network security company, and Alexandre Pauwels, a cybersecurity researcher at the University of Cambridge. In total, there appeared to be close to 3,000 assets linked to Anthropic's blog that had not been published previously on the company's news or research sites that were nonetheless publicly-accessible in this data cache, according to Pauwels, who Fortune asked to assess and review the material. After being informed of the data leak by Fortune on Thursday, Anthropic removed the public's ability to search the data store and retrieve documents from it. In a statement provided to Fortune, Anthropic acknowledged that a "human error" in the configuration of its content management system led the draft blog post to being accessible. It described the unpublished material that was left in an unsecured and publicly-searchable data store as "early drafts of content considered for publication." As well as referring to Mythos, the draft blog post also discussed a new tier of AI models that it says will be called "Capybara". In the document, Anthropic says: "'Capybara' is a new name for a new tier of model: larger and more intelligent than our Opus models -- which were, until now, our most powerful." Capybara and Mythos appear to refer to the same underlying model. Currently, Anthropic markets each of its models in three different sizes: the largest and most capable model versions are branded Opus, while a slightly faster and cheaper, but less capable, versions are branded Sonnet, and the smallest, cheapest, and fastest are called Haiku. However, in the blog post, Anthropic describes Capybara as a new tier of model that is even larger and more capable than Opus, but also more expensive. "Compared to our previous best model, Claude Opus 4.6, Capybara gets dramatically higher scores on tests of software coding, academic reasoning, and cybersecurity, among others," the company said in the blog. The document also said the company had completed training "Claude Mythos," which the draft blog post described as "by far the most powerful AI model we've ever developed." In response to questions about the draft blog post, the company acknowledged training and testing a new model. "We're developing a general purpose model with meaningful advances in reasoning, coding, and cybersecurity," an Anthropic spokesperson said. "Given the strength of its capabilities, we're being deliberate about how we release it. As is standard practice across the industry, we're working with a small group of early access customers to test the model. We consider this model a step change and the most capable we've built to date." The document Fortune and the cybersecurity experts reviewed consists of structured data for a webpage, complete with headings and a publication date, suggesting it forms part of a planned product launch. It outlines a cautious rollout strategy for the model, beginning with a small group of early-access users. The draft blog notes that the model is expensive to run and not yet ready for general release. The new AI model poses significant cybersecurity risks, according to the leaked document. "In preparing to release Claude Capybara, we want to act with extra caution and understand the risks it poses -- even beyond what we learn in our own testing. In particular, we want to understand the model's potential near-term risks in the realm of cybersecurity -- and share the results to help cyber defenders prepare," the document said. Anthropic appears to be especially worried about the model's cybersecurity implications, noting that the system is "currently far ahead of any other AI model in cyber capabilities" and "it presages an upcoming wave of models that can exploit vulnerabilities in ways that far outpace the efforts of defenders." In other words, Anthropic is concerned that hackers could use the model to run large-scale cyberattacks. The company said in the draft blog that because of this risk, its plan for the model's release would focus on cyber defenders: "We're releasing it in early access to organizations, giving them a head start in improving the robustness of their codebases against the impending wave of AI-driven exploits." The latest generation of frontier models from both Anthropic and OpenAI have crossed a threshold that the companies say poses new cybersecurity risks. In February, when OpenAI released GPT-5.3-Codex, the company said it was the first model it had classified as "high capability" for cybersecurity-related tasks under its Preparedness Framework -- and the first it had directly trained to identify software vulnerabilities. Anthropic, meanwhile, navigated similar risks with its Opus 4.6, released the same week. The model demonstrated an ability to surface previously unknown vulnerabilities in production codebases, a capability that the company acknowledged was dual-use, meaning that it could both help hackers as well as help cybersecurity defenders find and close vulnerabilities in code. The company has also reported that hacking groups, including those linked to the Chinese government, have attempted to exploit Claude in real-world cyberattacks. In one documented case, Anthropic discovered that a Chinese state-sponsored group had already been running a coordinated campaign using Claude Code to infiltrate roughly 30 organizations -- including tech companies, financial institutions, and government agencies -- before the company detected it. Over the following ten days, Anthropic investigated the full scope of the operation, banned the accounts involved, and notified affected organizations. The leak of not-yet-public information appears to stem from an error on the part of users of the company's content management system (CMS), which is the software used to publish the company's public blog, according to cybersecurity professionals. Digital assets created using the content management system are set to public by default and typically assigned a publicly accessible URL when uploaded -- unless the user explicitly changes a setting so that these assets are kept private. As a result, a large cache of images, PDF files, and audio files seem to have been published erroneously to an unsecured and publicly-accessible URL via the off-the-shelf content management system. Anthropic acknowledged in a statement to Fortune that "an issue with one of our external CMS tools led to draft content being accessible." It attributed this issue to "human error." Many of the documents appeared to be discarded or unused assets for past blog posts like images, banners, and logos. However, several appeared to be what were meant to be private or internal documents. For example, one asset has a title that described an employee's "parental leave." The documents also included a PDF containing information about an upcoming, invite-only retreat for the CEOs of European companies being held in the U.K., and which Anthropic CEO Dario Amodei will attend. Names of the other attendees are not listed, but are described as Europe's most influential business leaders. The two-day retreat is described as an "intimate gathering" to engage in "thoughtful conversation" at an 18th-century manor-turned-hotel-and-spa in the English countryside. The document says that attendees will hear from lawmakers and policymakers about how businesses are adopting AI and experience unreleased Claude capabilities. An Anthropic spokesperson told Fortune the event "is part of an ongoing series of events we've hosted over the past year. We look forward to hosting European business leaders to discuss the future of AI."
[7]
Anthropic's 'Most Capable' AI Model Claude Mythos Leaks, Deemed Major Cybersecurity Threat - Decrypt
Cybersecurity stocks declined after reports suggested the system could accelerate AI-driven cyberattacks. Claude creator Anthropic is developing a new AI model called Claude Mythos, described internally as the company's most capable model to date, with draft materials about the system being leaked online this week. The existence of the model was first reported by Fortune on Thursday after unpublished files tied to Anthropic's blog were discovered in a publicly accessible data cache. An Anthropic spokesperson confirmed the existence of the model to the publication. "We're developing a general purpose model with meaningful advances in reasoning, coding, and cybersecurity," an Anthropic spokesperson told Fortune. "Given the strength of its capabilities, we're being deliberate about how we release it. As is standard practice across the industry, we're working with a small group of early access customers to test the model. We consider this model a step change and the most capable we've built to date." In an archived development page reviewed by Decrypt, Anthropic called Mythos "the most powerful AI model we've ever developed." "Mythos is a new name for a new tier of model: larger and more intelligent than our Opus models -- which were, until now, our most powerful," Anthropic wrote. "We chose the name to evoke the deep connective tissues that link together knowledge and ideas." According to Anthropic, Mythos scored "dramatically higher" than Claude Opus 4.6 on tests of software coding, academic reasoning, and cybersecurity. The leak of Mythos appears to have originated from draft materials stored in an unsecured content management system. According to Fortune, Anthropic restricted public access to the data store after being notified that the files were searchable online. The company attributed the exposure to human error in the configuration of its CMS tools. However, Anthropic's documents labeled Mythos as version one of the new model, and described version two internally as "Capybara," which the company also positioned above its current top-tier Opus models. The draft materials also highlighted concerns about the system's potential cybersecurity implications. "Although Mythos is currently far ahead of any other AI model in cyber capabilities, it presages an upcoming wave of models that can exploit vulnerabilities in ways that far outpace the efforts of defenders," the company wrote. Because of those risks, the company said it plans to release the model cautiously, beginning with a limited early-access rollout aimed at organizations working on cybersecurity defense. Anthropic did not immediately respond to Decrypt's request for comment. While Anthropic took down the blog post, news of the leak quickly spilled into financial markets. Shares of several cybersecurity firms dropped after the reports surfaced, including Palo Alto Networks (PANW), which fell about 7%, and CrowdStrike (CRWD), which dropped roughly 6.4%. Meanwhile, Zscaler (ZS) declined around 5.8%, and Fortinet (FTNT) slipped about 4% during Friday trading, according to Yahoo Finance. The selloff reaction echoes a similar market response to the reveal of a new Anthropic product. In February, Anthropic unveiled Claude Cowork, an AI system designed to automate complex workplace tasks -- including contract review and compliance -- which triggered a broad sell-off across software and professional-services companies. That sell-off erased roughly $285 billion in market value as investors reassessed the long-term impact of AI agents on enterprise software businesses. "The market's response was a signal, not that AI agents will immediately replace these businesses, but that investors are finally pricing in the structural risk that foundation model providers can now compete directly with the software layer," Nexatech Ventures founder Scott Dylan told Decrypt at the time. "That's a polite way of saying if Anthropic can build a legal workflow tool in-house, what's stopping them from doing the same for finance, procurement, or HR?"
[8]
Why Anthropic's leaked AI model 'Mythos' poses cybersecurity risks
A human error in Anthropic's content management system accidentally exposed a draft blog that described the model. Anthropic is working on a new powerful artificial intelligence (AI) model that "poses unprecedented cybersecurity risks," according to a leak from the company. Last week, a data leak revealed that Anthropic was working on a new, powerful model, which the AI company has since acknowledged is called "Claude Mythos". An Anthropic spokesperson told Fortune it marks a "step change" in AI and it is the "most capable" model it has built so far with "meaningful advances in reasoning, coding, and cybersecurity". The technology could be so strong that it could be a hacker's dream device. Cybersecurity stocks slumped following the Anthropic rumours. Meanwhile, Anthropic is privately warning top government officials that Mythos makes large-scale cyberattacks much more likely in 2026, Axios reported. Here is everything we know about Mythos. A configuration error, or human error, in Anthropic's content management system accidentally exposed a draft blog that described the model. The blog, cited by Fortune, which was "available in an unsecured and publicly-searchable data store," said that the new model "poses unprecedented cybersecurity risks". Mythos is said to be a part of Capybara, which is a new tier of AI models t Fortune cited Antropic as saying in the document, "'Capybara' is a new name for a new tier of model: larger and more intelligent than our Opus models -- which were, until now, our most powerful." The model's advanced capabilities appear to suggest the AI systems can tackle more intricate problems with more autonomy and precision. Capybara and Mythos appear to refer to the same underlying model. As AI companies release improved models, AI agents- which are learning to act and reason without human input- get better too. Hackers can thereby run multiple hacking campaigns at once, which becomes more difficult to protect against. At the same time, employees are using AI agents which can unknowingly connect to their own work systems, which gives cybercriminals a door to enter. Additionally, human identities are now easier to breach thanks to AI, a Palo Alto Networks executive told Euronews Next in March. Mythos is extremely compute-intensive and expensive to run. Anthropic said it is working on making it much more efficient before any general release. Euronews Next has reached out to Anthropic for comment.
[9]
Anthropic tests Claude Mythos as its most powerful AI model yet
An AI model named Claude Mythos is reportedly being tested by Anthropic, following a data leak that suggests it is the company's most powerful model to date. The model is currently being trialed by a select group of early access customers. The development has raised concerns in the AI community as it follows OpenAI's halt of its own AI projects, including Sora, which has led to speculation about intensified competition in AI advancements. Experts highlight risks associated with jailbroken AI models, pointing to incidents where safety measures were stripped away, revealing alarming capabilities. Anthropic claims that Claude Mythos represents a significant advancement, with improved performance in reasoning, coding, and cybersecurity compared to prior models. An Anthropic spokesperson stated, "Given the strength of its capabilities, we're being deliberate about how we release it." This selective testing aligns with industry practice for managing potential risks. However, concerns are heightened due to a publicly available draft that indicated the model may introduce "unprecedented cybersecurity risks." Anthropic previously acknowledged uncertainties about the consciousness and moral status of their AI systems, further intensifying discussions about the implications of Claude Mythos. Social media reactions reveal a mix of excitement and skepticism. Some users noted that the language used to describe Mythos suggests a substantial leap in capability relative to previous releases. Others criticized the claims as exaggerated without tangible evidence. As the testing of Claude Mythos unfolds, the broader implications of its deployment remain under scrutiny. The potential risks associated with the model, alongside the unsettling history of hallucinations and harmful decisions linked to existing AI technologies, underscore the urgent need for careful oversight in AI development.
[10]
Claude Mythos: Leak spills details on Anthropic's new AI model, its most powerful yet - The Economic Times
A data leak revealed Anthropic is developing "Claude Mythos", its most powerful AI model yet, now in early testing. Exposed files showed details about the new models and cybersecurity risks that may result from it. The company blamed human error for the data leak.A data leak has revealed that Anthropic is developing a new artificial intelligence model it claims is its most powerful yet, with the system already being tested by a small group of users. A report in Fortune quoted an Anthropic spokesperson as saying the system is "the most capable we've built to date." Big leak Details about the model emerged after internal material was accidentally exposed in a public data store. In total, 3,000 assets linked to Anthropic's blog were accessible online. These included early drafts of announcements and other internal content that had not yet been released publicly. Among the files was a draft blog post referring to the model as "Claude Mythos" and warning that it could pose serious cybersecurity risks. The same leak also pointed to a planned, invite-only CEO summit in Europe, part of the company's push to promote its AI systems to large businesses. The company later said the leak had occurred due to "human error," specifically in how its content management system (CMS) was set up. It described the material as "early drafts of content considered for publication" and has since restricted access to the data. A new generation of AI models The leaked draft also referred to a new category of models under the name "Capybara." According to the document, this would represent a step beyond the company's current top-tier models. "'Capybara' is a new name for a new tier of model: larger and more intelligent than our Opus models -- which were, until now, our most powerful," Anthropic said in one leaked blogpost. Capybara and Mythos seem to be referring to the same underlying model, according to Fortune. Currently, Anthropic offers models at three levels: Opus, Sonnet and Haiku, which vary in size, cost and capability. Opus is the largest and most capable, designed for complex tasks but at a higher cost. Sonnet is a mid-tier option, balancing performance, speed and price. Haiku is the smallest, fastest and cheapest, suited for simpler use cases. The new system appears to go beyond Opus, making it both more advanced and more expensive. The document also suggested that training for "Claude Mythos" has already been completed. Cybersecurity concerns The leaked material highlights growing concern within the company about the risks linked to more advanced AI systems, Fortune said. "In preparing to release Claude Capybara, we want to act with extra caution and understand the risks it poses -- even beyond what we learn in our own testing. In particular, we want to understand the model's potential near-term risks in the realm of cybersecurity -- and share the results to help cyber defenders prepare," the document said, according to Fortune. In simple terms, Anthropic believes the model could be used to find and exploit weaknesses in software much faster than current tools. This raises the risk of more frequent and large-scale cyberattacks if such systems fall into the wrong hands. Because of this, the company plans to release the model carefully, starting with trusted organisations. "We're releasing it in early access to organisations, giving them a head start in improving the robustness of their codebases against the impending wave of AI-driven exploits," Anthropic said in the draft blog, according to Fortune. Real-world misuse already detected Anthropic has already seen attempts to misuse its AI systems. The company said hacking groups, including some linked to China, have tried to exploit its tools in real-world operations. In one case, a state-backed group used Claude Code in a coordinated effort targeting around 30 organisations, including technology companies, financial institutions and government bodies. Anthropic said it identified the activity, blocked the accounts involved and informed those affected within days. The incident underlines the wider challenge facing AI companies: building more powerful systems while trying to limit how they might be misused.
[11]
Anthropic Leak Reveals Claude Mythos Model and Cybersecurity Risks
Anthropic's recently leaked AI model, Claude Mythos, has drawn attention for its advanced capabilities and the ethical dilemmas it presents. According to Wes Roth, the model demonstrates exceptional performance in areas like cybersecurity, academic reasoning, and software development, but its ability to both identify and exploit vulnerabilities has raised serious concerns. The leak of over 3,000 internal files has provided a detailed look at the model's design, revealing not only its technical sophistication but also the potential risks it poses to industries that depend on secure systems. Gain insight into how Claude Mythos operates within cybersecurity, including its capacity to detect system weaknesses and the ethical challenges tied to its dual-use nature. Discover the broader implications of the leak, from shifts in public perception to its impact on Anthropic's position in the AI market. Additionally, understand the operational hurdles associated with the model's computational demands and how it compares to competing systems in the field. Claude Mythos represents a significant evolution in AI technology, building on the foundation of Anthropic's earlier models, such as Haiku, Sonnet and Opus. This new model is engineered to excel across multiple domains, offering a range of advanced capabilities: The dual nature of its cybersecurity capabilities, strengthening defenses while also exposing weaknesses, has positioned Claude Mythos as both a innovative tool and a source of ethical and practical concerns within the AI landscape. The details surrounding Claude Mythos came to light after a content management system (CMS) error inadvertently exposed over 3,000 internal files. These files included sensitive technical specifications, internal communications and strategic plans. While Anthropic acted swiftly to remove the documents and attributed the incident to human error, the leak has fueled speculation about its true origins. Some industry analysts suggest that the leak might have been a calculated strategy to generate buzz ahead of a major announcement, pointing to the timing and the nature of the exposed information. Others argue that the inclusion of unrelated documents indicates a genuine mistake rather than a deliberate move. Regardless of the intent, the leak has undeniably amplified interest in Claude Mythos, sparking discussions about its potential impact and the challenges it may bring. Here are more guides from our previous articles and guides related to Claude Mythos that you may find helpful. The cybersecurity implications of Claude Mythos are profound and multifaceted. On one hand, the model offers tools to fortify defenses against increasingly sophisticated cyber threats, allowing organizations to proactively identify and address vulnerabilities. On the other hand, its ability to exploit weaknesses faster than defenders can react raises serious concerns about potential misuse. Anthropic has acknowledged these risks and plans to provide early access to select organizations, allowing them to strengthen their defenses before the model becomes widely available. However, the potential for AI-driven cybersecurity threats has already sparked alarm among experts, who warn of the challenges in regulating such powerful technology. The dual-use nature of Claude Mythos underscores the need for robust oversight and ethical guidelines to prevent unintended consequences. The leak of Claude Mythos has already sent ripples through the market, particularly in the cybersecurity sector. Stocks of major firms like CrowdStrike and Palo Alto Networks experienced noticeable declines as investors reacted to the potential disruption posed by the model's advanced capabilities. Beyond market reactions, the high computational costs associated with operating Claude Mythos present significant economic challenges. The model requires substantial resources to function effectively, making it expensive to deploy at scale. This raises critical questions about its accessibility and long-term viability, particularly for smaller organizations that may struggle to afford its benefits. As a result, the economic implications of Claude Mythos extend beyond individual companies, influencing broader discussions about the widespread access of AI technology. Despite its promise, Claude Mythos faces several challenges that could impact its adoption and success: These uncertainties highlight the competitive and rapidly evolving nature of the AI industry. As companies race to develop and deploy innovative models, the ability to address these challenges will play a crucial role in determining the success of Claude Mythos. The circumstances surrounding the leak have led to widespread speculation about Anthropic's intentions. Some observers believe the incident could be part of a strategic effort to generate interest ahead of a planned CEO event or the company's anticipated IPO later in 2026. This theory is supported by the timing of the leak and the nature of the exposed documents, which have drawn significant attention to Claude Mythos. However, skeptics argue that the inclusion of unrelated files undermines this narrative, suggesting that the leak was more likely the result of human error. Regardless of the cause, the incident has undeniably heightened awareness of Claude Mythos, sparking discussions about its potential applications, risks and the broader implications for the AI industry. As Anthropic moves forward, its immediate priorities include refining Claude Mythos to improve its efficiency and reduce operational costs. The company is also navigating the complexities of preparing for an IPO, which could influence its strategy for rolling out the model. The broader AI landscape continues to evolve rapidly, with new models and technologies emerging at an unprecedented pace. In this context, Claude Mythos represents both an opportunity and a challenge. Its new capabilities highlight the immense potential of advanced AI, but they also underscore the need for careful management to mitigate risks and ensure responsible deployment. Anthropic's ability to address these challenges will not only shape the future of Claude Mythos but also influence the trajectory of AI development and its impact on industries worldwide. Disclosure: Some of our articles include affiliate links. If you buy something through one of these links, Geeky Gadgets may earn an affiliate commission. Learn about our Disclosure Policy.
[12]
Anthropic's Unreleased Claude Mythos Might Be The Most Advanced AI Model Yet | PYMNTS.com
By completing this form, you agree to receive marketing communications from PYMNTS and to the sharing of your information with our sponsor, if applicable, in accordance with our Privacy Policy and Terms and Conditions. According to a Thursday (March 26) report by Fortune, a configuration error in Anthropic's content management system left nearly 3,000 unpublished documents in a publicly searchable data store, including a draft blog post describing the model as "by far the most powerful AI model we've ever developed." Independent security researchers at LayerX Security and the University of Cambridge found the materials before Anthropic restricted access. An Anthropic spokesperson confirmed the model's existence to Fortune, calling it a step change and the most capable system the company has built, with meaningful advances in reasoning, coding and cybersecurity. The company said it is testing the model, known as Claude Mythos, with a small group of early-access customers and has not set a general release date, partly because it remains expensive to run at scale. The leaked draft described Mythos as part of a new model tier called Capybara, positioned above its current top-tier Opus models in both capability and cost. Where prior models respond to instructions one step at a time, Mythos plans and executes sequences of actions on its own, moving across systems, making decisions and completing operations without waiting for human input at each stage. As reported by Fortune, the leaked document described Mythos as currently far ahead of any other AI model in cybersecurity capabilities and said it signals an approaching generation of systems that can find and exploit software weaknesses faster than defenders can close them. Anthropic said its rollout plan prioritizes enterprise security teams, giving defenders early access before the model reaches wider distribution. According to a Sunday (March 29) report from Axios, Anthropic has been privately warning senior government officials that Mythos makes large-scale cyberattacks significantly more likely in 2026, and that agents running on systems at this capability level can plan and carry out complex operations with minimal human involvement. According to a separate Axios report, published in November, a Chinese state-sponsored hacking group in September used an earlier Claude model to carry out 80-90% of a coordinated attack campaign on its own, working through roughly 30 organizations including technology companies, financial institutions and government agencies before Anthropic detected and shut it down. The AI identified targets, found weaknesses, wrote attack code and produced detailed post-operation reports, all with minimal human direction. The operators running the attack convinced the model it was performing legitimate security testing. Once inside that framing, the AI executed the operation without further instruction. A Dark Reading poll published in January found that 48% of cybersecurity professionals now rank agentic AI as the top attack vector for 2026, above deepfakes and social engineering. As reported by PYMNTS, the September Claude Code incident marked the first confirmed case in which an AI agent handled most steps of a cyberattack normally performed by human operators. Eva Nahari, then-chief product officer at AI solutions firm Vectara, told PYMNTS the campaign was "global, industry-agnostic and growing," adding that with automation comes velocity and scale, and that attackers are now acquiring the same advantages that AI gives enterprises. As also reported by PYMNTS, Anthropic's earlier research found that its Claude Opus 4.5 model reduced successful prompt injection attacks to 1% in browser-based operations, down from higher breach rates in earlier versions, though the underlying vulnerability persists as browser-based automation grows more common. PYMNTS Intelligence found that 98% of business leaders remain unwilling to grant AI agents action-level access to core systems, with trust as the primary constraint on adoption. According to a Monday (March 30) report by CSO Online, shares of major cybersecurity vendors, including CrowdStrike, Palo Alto Networks, Zscaler and Fortinet, fell following the Mythos news as investors considered what frontier AI capabilities embedded in security tools could mean for the industry's competitive structure.
[13]
Meet Claude Mythos : Anthropic's Powerful Successor to Opus
Anthropic's Claude Mythos represents a new chapter in artificial intelligence, building on the foundation of the earlier Opus series. According to AI Grid, this model demonstrates strengths in areas such as academic reasoning, software development and cybersecurity. For instance, its ability to detect software vulnerabilities with precision underscores its potential to improve digital infrastructure security. At the same time, the model's high computational requirements and the ethical risks associated with misuse have led Anthropic to take a measured approach to its release and application. Explore how Claude Mythos tackles complex tasks like synthesizing knowledge across multiple domains and addressing coding challenges. Learn about the ethical measures Anthropic has put in place to address potential misuse, as well as the strategies being developed to manage the model's computational demands. This guide provides a closer look at how these elements shape the model's role in advancing responsible AI practices. Claude Mythos is designed to redefine the standards of AI performance by excelling in tasks that demand both precision and complexity. Its ability to synthesize knowledge across multiple domains enables it to identify software vulnerabilities, tackle intricate coding problems and contribute to academic research. These capabilities mark a significant leap forward, establishing Claude Mythos as a model that combines intelligence, scalability and practical applicability. By addressing real-world challenges, it sets a new benchmark for what AI systems can achieve in terms of both functionality and impact. The advanced capabilities of Claude Mythos are particularly evident in its proficiency in cybersecurity. It can detect vulnerabilities in systems with remarkable accuracy, offering a robust tool for safeguarding digital infrastructure. However, these same capabilities also raise ethical concerns. The potential misuse of such tools for malicious purposes, such as orchestrating cyberattacks, underscores the dual-use nature of advanced AI technologies. To mitigate these risks, Anthropic has implemented rigorous testing protocols and safeguards aimed at minimizing misuse. This approach aligns with a growing industry trend to address the ethical and security implications of deploying powerful AI systems, making sure that their benefits outweigh potential harms. Become an expert in Claude Mythos with the help of our in-depth articles and helpful guides. Despite its new potential, Claude Mythos faces several development challenges. One of the most pressing issues is its high computational demand, which translates into significant operational costs. These costs can limit accessibility and scalability, particularly for smaller organizations. To address this, Anthropic is exploring innovative techniques such as model distillation. This process involves creating smaller, more efficient versions of the model that retain its core capabilities. Enhancing computational efficiency is critical not only for reducing costs but also for broadening the model's accessibility, making sure that its benefits can be realized across a wider range of applications. The development of Claude Mythos has not been without setbacks. A notable incident involved a configuration error in Anthropic's content management system, which inadvertently leaked early details about the model. This breach has sparked discussions about the security measures surrounding sensitive AI projects and the potential consequences of such lapses. While some have speculated that the leak could have been a calculated publicity move, Anthropic has reaffirmed its commitment to maintaining stringent security protocols. This incident highlights the importance of robust safeguards in protecting intellectual property and making sure the integrity of AI development processes. Claude Mythos is primarily tailored for enterprise-level users, reflecting its advanced capabilities and the associated operational costs. This focus on large-scale applications raises questions about its accessibility for smaller organizations and individual users. The pricing structure is expected to align with its premium features, potentially limiting its adoption to well-funded enterprises. Despite these limitations, the model's potential to transform industries such as finance, healthcare and cybersecurity remains substantial. Its ability to address complex, real-world challenges positions it as a fantastic tool for organizations seeking to use innovative AI solutions. The release of Claude Mythos signals a new era in artificial intelligence, with the potential to reshape the technological landscape. Its ability to self-improve and tackle complex challenges positions it as a key player in the next generation of AI systems. However, its long-term success will depend on Anthropic's ability to address critical issues such as security vulnerabilities and computational efficiency. If these challenges are effectively managed, Claude Mythos could pave the way for further advancements in AI, influencing both technological innovation and ethical considerations. Its development serves as a reminder of the delicate balance between innovation and responsibility in the rapidly evolving field of artificial intelligence. Following the leak of early details about Claude Mythos, speculation has emerged regarding a secondary model within the Mythos tier, tentatively named Claude Capiara. This rumored model is believed to serve as a complementary version of Claude Mythos, similar to the sub-models introduced in the Opus series. While Anthropic has not officially confirmed its existence, the possibility of a secondary model highlights the company's commitment to expanding its AI ecosystem. If realized, Claude Capiara could offer additional functionalities or cater to specific use cases, further solidifying Anthropic's position in the competitive AI market. Claude Mythos enters a highly competitive AI market dominated by other advanced models such as GPT-4 and Gemini. Comparisons among these systems underscore the rapid pace of innovation and the growing demand for sophisticated AI solutions. Anthropic's emphasis on continuous improvement and rigorous internal testing suggests that Claude Mythos is not only a response to current market demands but also a precursor to future developments within its ecosystem. Its release is likely to influence the trajectory of AI innovation, setting new standards for performance, security and ethical considerations in the field. Disclosure: Some of our articles include affiliate links. If you buy something through one of these links, Geeky Gadgets may earn an affiliate commission. Learn about our Disclosure Policy.
[14]
Claude Mythos - A New Model that Anthropic was Forced to Reveal After a Data Leak
Did Anthropic's new foundational model reveal itself to the outside world? Or are we just becoming paranoid? Anthropic may be battling an entire political administration, but their researchers are on the ball when it comes to developing new foundational models that challenge competition. However, this time the reveal happened earlier than they'd have wanted. All because of a data leak caused by Anthropic's public-facing content management system. Speculative reports appeared first on The Fortune last Friday where the publication revealed that Anthropic had "inadvertently revealed details of an upcoming model release, an exclusive CEO event and some internal data, images and PDFs. Immediately thereafter, the publication got confirmation from Anthropic that it was testing a new AI model that could be its best-ever. In fact, the publication quoted an Anthropic spokesperson to state that the new model was already under testing with early access customers and may represent a "step change" in AI performance and could emerge "as the most capable we've built till date." From Anthropic's point of view, this could be a major challenge for the company, given the amount of media coverage that they received for their Claude Cowork and Claude Code - especially since these two tools seem to have raised pertinent questions around the continuance of the software-as-a-solution model of tech business. Based on the reports by Fortune, the leaks also point to a new tier of AI models that have been called Capybara. In the document, Anthropic reportedly says: "'Capybara' is a new name for a new tier of model: larger and more intelligent than our Opus models -- which were, until now, our most powerful." Now Claude Mythos appears to be part of this tier, though there is no clarity around how Capybara fits into Anthropic's existing tiers comprising Opus, Sonnet, and Haiku. Especially in their reduction in overall size, increase in capability and slashing overall costs. "Compared to our previous best model, Claude Opus 4.6, Capybara gets dramatically higher scores on tests of software coding, academic reasoning, and cybersecurity, among others," says a direct quote ascribed to the now missing blog post by Fortune. Meanwhile, the spokesperson also noted that the new model, called Claude Mythos, was a general purpose one with "meaningful advances in reasoning, coding, and cybersecurity. Of course, the irony was not lost on the readers as it was a draft blog available in "an unsecured and publicly searchable data store" that gave out first details of the model. What was worse is that the draft blog claimed that the new model could pose "unprecedented cybersecurity risks", which set everyone into a tizzy as some interpreted it as an acceptance that the new model itself could have leaked the information! The article by Fortune said the Anthropic left the material announcing a new model in an unsecured, public data lake. It said this information came from a separate set of documents separately located and verified by AI security researcher Roy Paz of LayerX Security and Alexandre Pauwels, a cybersecurity researcher at the University of Cambridge. The report cited more than 3,000 assets linked to Anthropic's blog that had not been published previously on the company's news and research sites. Later, Anthropic acknowledged it as a "human error" in a statement sent only to Fortune. It described the unpublished material as "early drafts of content considered for publication." Whatever be the issue with the leaks, the reveal has left the AI space more confused, especially with reference to the Capybara. For long, Anthropic has marketed each of its model in three different sizes - Opus which is faster and cheaper, Sonnet that is less capable and Haiku that is smallest, cheapest and fastest. Does it mean Capybara is a new tier to this structure? "Compared to our previous best model, Claude Opus 4.6, Capybara gets dramatically higher scores on tests of software coding, academic reasoning, and cybersecurity, among others," the company reportedly said in the blog. It also noted that Anthropic had completed training Claude Mythos, which the blog had described as "by far the most powerful AI model we've ever developed." Later, the article in Fortune quoted the spokesperson to state that Anthropic was developing a general purpose model with meaningful advances in reasoning, coding and cybersecurity. "Given the strength of its capabilities, we're being deliberate about how we release it. As is standard practice across the industry, we're working with a small group of early access customers to test the model. We consider this model a step change and the most capable we've built to date," the magazine quoted the spokesperson as saying. However, the leaked documents also outlines a rollout strategy for the model starting with a small group of early-access users. However, the draft blog notes at another place that the model is expensive to run and not yet ready for general release. Looks like Anthropic has landed itself in a bit of a mess that could take the sheen off the company's reputation as a cybersecurity-led software development space. Not to mention the stories around how the leadership team's boasts about automating much of its internal software development. We hope, for their sake, that AI did not leak its own prowess.
[15]
Claude Mythos Leak Sparks Alarm Over AI-Driven Cyber Threats
Anthropic Claude Mythos leak raises fears of AI-powered cyberattacks and security risks Reports about a model dubbed 'Claude Mythos' have raised fresh alarms over artificial intelligence and cyber threats. The latest claims link the system to Anthropic's Claude and suggest it can identify vulnerabilities and generate exploit code. Anthropic has not confirmed any such model. Current reports rely on interpretations of recent security findings, not an official release. Researchers have already shown that AI tools can assist in cyberattacks. Hackers can use these systems to scan code, detect flaws, and build attack scripts. These tools cut the skill needed to launch sophisticated attacks. Security analysts have also found weaknesses in , which could expose systems to misuse. There are three risks identified by experts in the immediate future. First, attackers can misuse AI tools with ease. The growth of cybercrimes is faster than the . There is also a possibility for systems to behave in an unpredictable manner with less human intervention. Advanced AI can be used for coordinated attacks with little supervision. The leak is not confirmed, but the risks are real. AI is advancing faster than security frameworks. The focus now is on managing active threats, not speculating on new models.
[16]
Anthropic's Claude leaks complicate its responsible AI narrative
Operational missteps raise questions about AI stewardship credibility Anthropic couldn't Claude its way out of the hole it has dug for itself over the past week or so any faster. After topping the app store as the #1 AI chatbot ahead of ChatGPT and Gemini, following its moral war with the US Department of Defense, Anthropic, the steward of responsible AI, has behaved irresponsibly to such an extent it beggars belief. Anthropic earned continued praise for publicly releasing its anti-retaliation process around Responsible Scaling Policy concerns, which allows employees to report AI safety issues without facing any retribution. The Claude maker then got hit by internal data exposure and a much bigger Claude Code source leak within days. All of this happened within a week, the pace itself dizzying. I'm imagining people who have been applauding Anthropic's continued effort to reinforce its brand around responsible AI and AI safety hadn't even stopped clapping, before being left to scratch their head in utter confusion and raw disbelief. After its public face-off with the US Department of Defense, where Anthropic CEO Dario Amodei drew a line in the sand on allowing Claude to be used for fully autonomous weapons systems, the AI company kept presenting itself as the industry's responsible AI poster child with lots of key announcements. On March 11, the Claude maker launched Anthropic Institute to share its perspectives on how society should navigate difficult decisions around increasingly more powerful AI systems. Just before that, it had released a transparency hub on responsible AI development dos and don'ts for the industry. On March 24, it dropped the responsible scaling policy to assure and encourage employees of raising retribution-free internal alarms on AI developments. Anthropic also signed an AI safety partnership with the Australian government on April 1 to signal how nation states should deal with matters related to responsible AI. So far so good. On March 26, reports showed how Anthropic had accidentally leaked Claude Mythos, an unreleased AI model that it claimed could cause havoc online in terms of cybersecurity. It all came to light by a CMS configuration error, according to Anthropic, which exposed hundreds of unpublished documents, including details of an invite-only CEO event to prepare the industry on Claude Mythos' cybersecurity implications. Then on March 31, a release packaging error pushed an unwanted file into the public facing library of Claude Code. What did it contain? Links to nearly 2,000 files and approximately 500,000 lines of source code of Claude Code itself! No user data, private credentials or sensitive weights of the Claude Code model were compromised, according to Anthropic, but that was hardly the point. These instances were body blows to Anthropic's responsible AI, safety-first lab image, in the aftermath of having shipped its own internal secrets to the public. How can Anthropic advise the world at large on how to build responsible AI, while it struggles with basic discipline needed to handle sensitive material through its own systems? Yeah, sure, the argument can be made that Anthropic's leaks aren't in the same category of failure as releasing a reckless frontier model into the open. Anthropic may still be more serious than many rivals about AI safety as a philosophy, but recent history has shown it lacks the basic ability to practice what it preaches so loudly about AI safety. An unwanted, self-inflicted wound that tarnishes some of the glitter Anthropic's brand had accrued recently.
[17]
Anthropic's Claude Mythos leak and what it means for AI and cybersecurity
Mythos presents an important inflection point in AI and cybersecurity Anthropic is quickly earning a reputation for being a responsible AI standard bearer, especially after CEO Dario Amodei's moral clash with the US Department of Defense in Feb 2026. But the fact that its upcoming model, dubbed Claude Mythos, leaked online ahead of schedule thanks to a CMS error is ironic to say the least. According to the leaked draft of Anthropic's blog post, Claude Mythos is so powerful and "far ahead of any other AI model in cyber capabilities" that it can lead to an overwhelming "wave of models that can exploit vulnerabilities in ways that far outpace the efforts of defenders." That's Anthropic suggesting they're cooking up something so powerful in their next AI product it can essentially wreak havoc online in terms of cybersecurity, something that the world's best cybersecurity companies won't be able to handle - with or without AI. That's what scares Anthropic about Mythos' capabilities, and why (according to the leaked documents) they were planning to engage top business leaders behind closed doors and brief and prepare them on next steps. How do we know all this ahead of schedule? Because someone at Anthropic accidentally misconfigured these posts on the website's CMS, making thousands of documents publicly viewable. One wonders if it isn't a euphemistic rephrasing of a cyber breach. Irony-laced jokes about Claude Mythos notwithstanding, Anthropic claims the Mythos model is currently being trialed by "early access customers" only, deliberately and carefully rolling out much slower than previous Claude models. Also read: Anthropic accidentally reveals Claude Mythos, its most powerful AI model yet That underlying concern Anthropic's exhibiting isn't just limited to them. AI has the potential to disrupt the cybersecurity industry rapidly, for better and worse. There's data to back it up. According to Crowdstrike, AI-enabled attacks jumped 89% year-over-year in 2025, noting that the speed and scale of cyber breaches was only accelerating. Business email hack attacks rose 37% in 2025, suggests a Deepstrike report, using AI deepfake tech like voice cloning and synthetic identities. In its 2026 AI security report, Cisco has warned that AI is fast expanding the cyber attack surface with potential new risks like agent misuse and supply-chain attacks. In fact, cybercriminals are "weaponizing AI to orchestrate attacks of unprecedented sophistication and scale," according to Kaspersky Lab's Evolving Threat Landscape Report of December 2025. According to the cybersecurity company, "malicious actors are deploying generative AI to create convincing phishing emails, voice clones, and deepfake videos that bypass traditional detection mechanisms." These AI-driven attacks will target AI models corporate systems through evasion, poisoning, and model extraction techniques, said Kaspersky. Claude Mythos, still unreleased and already controversial, seemingly sits at the crossroads of AI-linked cybersecurity. Because, from the looks of it, the same AI model powerful enough to identify zero-day vulnerabilities and fortify codebases is also equally capable of being weaponised in the wrong hands. Anthropic's decision to initially limit access to defensive cybersecurity use cases is a measured response, but only for the time being. It can't control what others working the AI crucible will eventually forge and unleash.
[18]
Anthropic accidentally reveals Claude Mythos, its most powerful AI model yet
The company is taking a cautious rollout approach due to fears the model could be misused for large-scale cyberattacks, echoing concerns also seen at OpenAI. Anthropic is currently developing a new AI model that can easily outperform current offerings, even as an internal data leak unintentionally revealed details about the project. The company has confirmed the model is being tested with a limited group of early access users. As per the information that surfaced from the publicly accessible internal files, the upcoming model may be called Claude Mythos. In a draft document reviewed by external sources, the model was described as the most powerful AI the company has made so far, with some improvements in reasoning, coding and cybersecurity capabilities. On the other hand, Anthropic acknowledged that exposure of these documents was due to a configuration error in the content management system, which made unpublished materials visible online. The company stated that the files were early drafts intended for internal use and has since restricted access to the data. The leaked material also pointed to a new classification of AI systems internally referred to as Capybara, which appears to represent a tier above existing models like Opus. Anthropic currently categorises its models into tiers such as Opus, Sonnet and Haiku, but the new system is expected to exceed these in both capability and cost. Also read: Google releases Lyria 3 Pro AI model with longer music generation: How to access The company has flagged potential cybersecurity risks associated with the model. The internal description suggests that its advanced features can be misused to find and exploit different software vulnerabilities at scale, which might outpace current defensive systems. As a result, Anthropic is said to be approaching the rollout cautiously, initially offering access to a few organisations to strengthen the cyber defense. Adding on, the model details that surfaced online also reportedly included references to a private executive event in Europe aimed at engaging business leaders on AI adoption. This comes amid the increasingly capable models raising concerns around dual-use risk. Recently, OpenAI has faced similar concerns, particularly around models designed to detect vulnerabilities in software systems.
Share
Share
Copy Link
Anthropic confirmed its most powerful AI model yet after an embarrassing data leak exposed details about Claude Mythos. The company warns the model presents unprecedented cybersecurity risks and could enable large-scale cyberattacks that far outpace defenders' capabilities. The leak comes as Anthropic eyes an IPO and faces Pentagon scrutiny.
Anthropic has confirmed the existence of Claude Mythos, what it describes as a step change in AI performance and "the most capable we've built to date," following an internal data leak that exposed nearly 3,000 unpublished assets
1
. The leak occurred due to human error in the company's content management system, which failed to mark sensitive materials as private, leaving them accessible in a publicly searchable data store5
. Fortune reporter Bea Nolan first identified the exposed data, which included an unpublished blog post describing the powerful new AI model as "by far the most powerful AI model we've ever developed"3
.
Source: Digit
The timing of this security lapse is particularly ironic given that the leaked materials themselves warned about unprecedented cybersecurity risks posed by Claude Mythos
4
. Anthropic's spokesperson confirmed to Fortune that the model is "a general purpose model with meaningful advances in reasoning, coding, and cybersecurity," though it remains in early access with select customers5
.The leaked draft blog post revealed that Anthropic is privately warning government officials that Claude Mythos makes large-scale cyberattacks much more likely in 2026
2
. According to the company's own assessment, the model is "currently far ahead of any other AI model in cyber capabilities" and "presages an upcoming wave of models that can exploit vulnerabilities in ways that far outpace the efforts of defenders"3
.The model allows AI agents to work autonomously with sophisticated precision to penetrate corporate, government, and municipal systems—essentially a hacker's dream weapon
2
. One source briefed on the coming models told Axios that a large-scale attack could hit this year, with businesses as ripe targets2
. The threat is amplified by employees testing agents without realizing they're creating entry points for cybercriminals through what the industry calls "shadow AI"2
.
Source: Axios
The leaked materials also revealed a new AI model tier called Capybara, which would sit above Anthropic's current top-tier Opus model
3
. Claude Mythos appears to be the first model in this new tier, with the leaked blog post stating that "Compared to our previous best model, Claude Opus 4.6, Capybara gets dramatically higher scores on tests of software coding, academic reasoning, and cybersecurity, among others"4
. Currently, Anthropic offers three tier levels: Opus as the most powerful and expensive, followed by Sonnet and Haiku3
.The model is reportedly very expensive to train and operate, so expensive that Anthropic isn't sure how to deploy it at this point
1
. The leaked draft indicated that the system's high running costs mean a public launch has not yet been scheduled5
.Related Stories
The revelation has given ammunition to the Pentagon, which has been locked in conflict with Anthropic after the company refused to allow its models to be used for domestic surveillance or fully autonomous military weapons
1
. Under Secretary of War Emil Michael, who has significant financial ties to Anthropic competitors, posted "Umm…hello? Is it not clear yet that we have a problem here?" following the leak1
. Michael has been calling CEO Dario Amodei a "liar" with a "god complex" for weeks1
.A judge recently handed down a ruling that temporarily blocked the Department of Defense from labeling Anthropic a security risk, though the Pentagon continues fighting in the court of public opinion
1
. Late last year, Anthropic disclosed the first documented case of a cyberattack largely executed by AI—a Chinese state-sponsored group that used AI agents to autonomously hack roughly 30 global targets, with the AI handling 80-90% of tactical operations independently2
.
Source: Gizmodo
Anthropic stated in the leaked blog that it was providing early access to organizations to give them "a head start in improving the robustness of their codebases against the impending wave of AI-driven exploits"
3
. A Dark Reading poll found that 48% of cybersecurity professionals now rank agentic AI as the #1 attack vector for 2026—above deepfakes and everything else2
.The new models enable bad actors to scale simply with more compute, no longer limited by finite personnel—like "a warehouse full of the most sophisticated criminals who never sleep, learn on the fly and persist until successful—except the warehouse is infinite"
2
. Cybersecurity stocks plunged on Friday following the news4
.The leak also exposed details about a planned private summit for European business leaders at a U.K. country manor, which Dario Amodei is set to attend
5
. The timing coincides with reports that Anthropic is eyeing an IPO later this year, making the revelation about this powerful new AI model potentially valuable information for prospective investors1
.Summarized by
Navi
[1]
Yesterday•Technology

23 May 2025•Technology

23 May 2025•Technology

1
Technology

2
Policy and Regulation

3
Policy and Regulation
