Cybersecurity

Anthropic locked down its most powerful AI Model over cybersecurity fears–then put it to work

Published byAIDaily Editorial Team
4 min read
Original source author: Dashveenjit Kaur

Anthropic’s most capable AI model has already found thousands of AI cybersecurity vulnerabilities across every major operating system and web browser. The company’s response was not to release it, but to quietly hand it to the organisations responsible for keeping the internet running. That model is Claude Mythos Preview, and the initiative is called Project Glasswing. […] The post Anthropic locked down its most powerful AI Model over cybersecurity fears–then put it to work appeared first on AI

Share:

Anthropic’s most capable AI model has already found thousands of AI cybersecurity vulnerabilities across every major operating system and web browser. The company’s response was not to release it, but to quietly hand it to the organisations responsible for keeping the internet running. That model is Claude Mythos Preview, and the initiative is called Project Glasswing . The launch partners include Amazon Web Services, Apple, Broadcom, Cisco, CrowdStrike, Google, JPMorganChase, the Linux Foundation, Microsoft, Nvidia, and Palo Alto Networks. Beyond that core group, Anthropic has extended access to over 40 additional organisations that build or maintain critical software infrastructure. Anthropic is committing up to US$100 million in usage credits for Mythos Preview across the effort, along with US$4 million in direct donations to open-source security organisations. A model that outgrew its own benchmarks Mythos Preview was not specifically trained for cybersecurity work. Anthropic said the capabilities “emerged as a downstream consequence of general improvements in code, reasoning, and autonomy”, and that the same improvements making the model better at patching vulnerabilities also make it better at exploiting them. That last part matters. Mythos Preview has improved to the extent that it mostly saturates existing security benchmarks, forcing Anthropic to shift its focus to novel real-world tasks–specifically, zero-day vulnerabilities. These flaws were previously unknown to the software’s developers. Among the findings: a 27-year-old bug in OpenBSD, an operating system known for its strong security posture. In another case, the model fully autonomously identified and exploited a 17-year-old remote code execution vulnerability in FreeBSD–CVE-2026-4747–that allows an unauthenticated user anywhere on the internet to obtain complete control of a server running NFS. No human was involved in the discovery or exploitation after the initial prompt to find the bug. Nicholas Carlini from Anthropic’s research team described the model’s ability to chain together vulnerabilities: “This model can create exploits out of three, four, or sometimes five vulnerabilities that in sequence give you some kind of very sophisticated end outcome. I’ve found more bugs in the last couple of weeks than I found in the rest of my life combined.” Why is it not being released? “We do not plan to make Claude Mythos Preview generally available due to its cybersecurity capabilities,” Newton Cheng, Frontier Red Team Cyber Lead at Anthropic, said. “Given the rate of AI progress, it will not be long before such capabilities proliferate, potentially beyond actors who are committed to deploying them safely. The fallout–for economies, public safety, and national security–could be severe.” This is not hypothetical. Anthropic had previously disclosed what it described as the first documented case of a cyberattack largely executed by AI–a Chinese state-sponsored group that used AI agents to autonomously infiltrate roughly 30 global targets, with AI handling the majority of tactical operations independently. The company has also privately briefed senior US government officials on Mythos Preview’s full capabilities. The intelligence community is now actively weighing how the model could reshape both offensive and defensive hacking operations. The open-source problem One dimension of Project Glasswing that goes beyond the headline coalition: open-source software. Jim Zemlin, CEO of the Linux Foundation, put it plainly: “In the past, security expertise has been a luxury reserved for organisations with large security teams. Open-source maintainers, whose software underpins much of the world’s critical infrastructure, have historically been left to figure out security on their own.” Anthropic has donated US$2.5 million to Alpha-Omega and OpenSSF through the Linux Foundation, and US$1.5 million to the Apache Software Foundation–giving maintainers of critical open-source codebases access to AI cybersecurity vulnerability scanning at a scale that was previously out of reach. What comes next Anthropic says its eventual goal is to deploy Mythos-class models at scale, but only when new safeguards are in place. The company plans to launch new safeguards with an upcoming Claude Opus model first, allowing it to refine them with a model that does not pose the same level of risk as Mythos Preview. The competitive picture is already shifting around it. When OpenAI released GPT-5.3-Codex in February, the company called it the first model it had classified as high-capability for cybersecurity tasks under its Preparedness Framework. Anthropic’s move with Glasswing signals that the frontier labs see controlled deployment–not open release–as the emerging standard for models at this capability level. Whether that standard holds as these capabilities spread further is, at this point, an open question that no single initiative can answer. See Also: Anthropic’s refusal to arm AI is exactly why the UK wants it Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and is co-located with other leading technology events including the Cyber Security & Cloud Expo . Click here for more information. AI News is powered by TechForge Media . Explore other upcoming enterprise technology events and webinars here . The post Anthropic locked down its most powerful AI Model over cybersecurity fears–then put it to work appeared first on AI News .

Key takeaways

  • Anthropic chose not to release the Claude Mythos Preview model due to its cybersecurity capabilities, raising ethical questions.
  • Collaboration with major tech companies highlights the importance of strategic partnerships in developing security solutions.
  • Brazil must keep up with developments in cybersecurity and AI to ensure it does not fall behind in technological innovation.

Editorial analysis

Anthropic's decision not to release the Claude Mythos Preview model, despite its impressive capabilities in identifying cybersecurity vulnerabilities, raises crucial questions about ethics and responsibility in AI development. For the Brazilian tech sector, which is growing and seeking to position itself as an innovation hub, this approach may serve as a warning about the need for regulation and responsible practices in the use of artificial intelligence. Data protection and cybersecurity are global concerns, and Brazil, with its legislation like the LGPD, must be vigilant about how these technologies can impact citizens' privacy and security.

Moreover, Anthropic's collaboration with major tech players like Amazon and Google highlights the importance of strategic partnerships in developing security solutions. This could inspire Brazilian companies to seek similar collaborations, fostering a more robust and secure ecosystem. The Project Glasswing initiative can also be seen as a model to follow, where social responsibility and security are prioritized over immediate profit.

Looking ahead, it is crucial for the tech sector in Brazil to closely monitor developments in cybersecurity and AI. The model's ability to identify and exploit vulnerabilities may lead to increased demand for information security professionals, creating new job opportunities. Additionally, the need to develop AI solutions that are ethical and secure should be a priority for startups and established companies in the country, ensuring that Brazil does not fall behind in the global race for technological innovation.

What this coverage includes

  • Clear source attribution and link to the original publication.
  • Editorial framing about relevance, impact, and likely next developments.
  • Review for readability, context, and duplication before publication.

Original source:

AI News

About this article

This article was curated and published by AIDaily as part of our editorial coverage of artificial intelligence developments. The content is based on the original source cited below, enriched with editorial context and analysis. Automated tools may assist with translation and initial structuring, but publication decisions, factual review, and contextual framing remain editorial responsibilities.

Learn more about our editorial process