Cybersecurity

Anthropic keeps new AI model private after it finds thousands of external vulnerabilities

Published byAIDaily Editorial Team
4 min read
Original source author: Dashveenjit Kaur

Anthropic’s most capable AI model has already found thousands of AI cybersecurity vulnerabilities across every major operating system and web browser. The company’s response was not to release it, but to quietly hand it to the organisations responsible for keeping the internet running. That model is Claude Mythos Preview, and the initiative is called Project Glasswing. […] The post Anthropic keeps new AI model private after it finds thousands of external vulnerabilities appeared first on AI News

Share:

Anthropic’s most capable AI model has already found thousands of AI cybersecurity vulnerabilities across every major operating system and web browser. The company’s response was not to release it, but to quietly hand it to the organisations responsible for keeping the internet running. That model is Claude Mythos Preview, and the initiative is called Project Glasswing . The launch partners include Amazon Web Services, Apple, Broadcom, Cisco, CrowdStrike, Google, JPMorganChase, the Linux Foundation, Microsoft, Nvidia, and Palo Alto Networks. Beyond that core group, Anthropic has extended access to over 40 additional organisations that build or maintain critical software infrastructure. Anthropic is committing up to US$100 million in usage credits for Mythos Preview across the effort, along with US$4 million in direct donations to open-source security organisations. A model that outgrew its own benchmarks Mythos Preview was not specifically trained for cybersecurity work. Anthropic said the capabilities “emerged as a downstream consequence of general improvements in code, reasoning, and autonomy”, and that the same improvements making the model better at patching vulnerabilities also make it better at exploiting them. That last part matters. Mythos Preview has improved to the extent that it mostly saturates existing security benchmarks, forcing Anthropic to shift its focus to novel real-world tasks–specifically, zero-day vulnerabilities. These flaws were previously unknown to the software’s developers. Among the findings: a 27-year-old bug in OpenBSD, an operating system known for its strong security posture. In another case, the model fully autonomously identified and exploited a 17-year-old remote code execution vulnerability in FreeBSD–CVE-2026-4747–that allows an unauthenticated user anywhere on the internet to obtain complete control of a server running NFS. No human was involved in the discovery or exploitation after the initial prompt to find the bug. Nicholas Carlini from Anthropic’s research team described the model’s ability to chain together vulnerabilities: “This model can create exploits out of three, four, or sometimes five vulnerabilities that in sequence give you some kind of very sophisticated end outcome. I’ve found more bugs in the last couple of weeks than I found in the rest of my life combined.” Why is it not being released? “We do not plan to make Claude Mythos Preview generally available due to its cybersecurity capabilities,” Newton Cheng, Frontier Red Team Cyber Lead at Anthropic, said. “Given the rate of AI progress, it will not be long before such capabilities proliferate, potentially beyond actors who are committed to deploying them safely. The fallout–for economies, public safety, and national security–could be severe.” This is not hypothetical. Anthropic had previously disclosed what it described as the first documented case of a cyberattack largely executed by AI–a Chinese state-sponsored group that used AI agents to autonomously infiltrate roughly 30 global targets, with AI handling the majority of tactical operations independently. The company has also privately briefed senior US government officials on Mythos Preview’s full capabilities. The intelligence community is now actively weighing how the model could reshape both offensive and defensive hacking operations. The open-source problem One dimension of Project Glasswing that goes beyond the headline coalition: open-source software. Jim Zemlin, CEO of the Linux Foundation, put it plainly: “In the past, security expertise has been a luxury reserved for organisations with large security teams. Open-source maintainers, whose software underpins much of the world’s critical infrastructure, have historically been left to figure out security on their own.” Anthropic has donated US$2.5 million to Alpha-Omega and OpenSSF through the Linux Foundation, and US$1.5 million to the Apache Software Foundation–giving maintainers of critical open-source codebases access to AI cybersecurity vulnerability scanning at a scale that was previously out of reach. What comes next Anthropic says its eventual goal is to deploy Mythos-class models at scale, but only when new safeguards are in place. The company plans to launch new safeguards with an upcoming Claude Opus model first, allowing it to refine them with a model that does not pose the same level of risk as Mythos Preview. The competitive picture is already shifting around it. When OpenAI released GPT-5.3-Codex in February, the company called it the first model it had classified as high-capability for cybersecurity tasks under its Preparedness Framework. Anthropic’s move with Glasswing signals that the frontier labs see controlled deployment–not open release–as the emerging standard for models at this capability level. Whether that standard holds as these capabilities spread further is, at this point, an open question that no single initiative can answer. See Also: Anthropic’s refusal to arm AI is exactly why the UK wants it Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and is co-located with other leading technology events including the Cyber Security & Cloud Expo . Click here for more information. AI News is powered by TechForge Media . Explore other upcoming enterprise technology events and webinars here . The post Anthropic keeps new AI model private after it finds thousands of external vulnerabilities appeared first on AI News .

Key takeaways

  • Anthropic chose not to release Claude Mythos Preview due to its cybersecurity capabilities.
  • The model demonstrated the ability to identify and exploit vulnerabilities, raising concerns about the malicious use of AI.
  • Collaboration with major companies may inspire cybersecurity initiatives in Brazil.

Editorial analysis

Anthropic's decision not to release Claude Mythos Preview, even after discovering thousands of cybersecurity vulnerabilities, raises crucial questions about responsibility in AI development. For the Brazilian tech sector, this serves as a warning about the need for a greater focus on cybersecurity, especially with the increasing adoption of AI across various applications. The collaboration with major tech companies and security organizations indicates a shift towards a more collaborative approach to risk mitigation, which could inspire similar initiatives in Brazil.

Moreover, the model's ability to identify and exploit vulnerabilities highlights the duality of AI: while it can be a powerful tool for security, it can also be used for malicious purposes. This suggests that Brazil should consider implementing ethical and regulatory guidelines for AI use, particularly in sensitive areas like cybersecurity. Developing policies that encourage transparency and accountability in the use of emerging technologies will be crucial to ensure that technological advancement does not compromise national security.

Finally, the Project Glasswing initiative could serve as a model for future collaborations between tech companies and security institutions in Brazil. Anthropic's significant investment in usage credits and donations to open-source security organizations may stimulate a more robust and secure ecosystem. We will watch how this collaborative approach influences the development of cybersecurity solutions in Brazil, especially at a time when digitalization is accelerating across all sectors of the economy.

What this coverage includes

  • Clear source attribution and link to the original publication.
  • Editorial framing about relevance, impact, and likely next developments.
  • Review for readability, context, and duplication before publication.

Original source:

AI News

About this article

This article was curated and published by AIDaily as part of our editorial coverage of artificial intelligence developments. The content is based on the original source cited below, enriched with editorial context and analysis. Automated tools may assist with translation and initial structuring, but publication decisions, factual review, and contextual framing remain editorial responsibilities.

Learn more about our editorial process