Segurança Cibernética

Anthropic keeps new AI model private after it finds thousands of external vulnerabilities

Publicado porRedacao AIDaily
4 min de leitura
Autor na fonte original: Dashveenjit Kaur

Anthropic’s most capable AI model has already found thousands of AI cybersecurity vulnerabilities across every major operating system and web browser. The company’s response was not to release it, but to quietly hand it to the organisations responsible for keeping the internet running. That model is Claude Mythos Preview, and the initiative is called Project Glasswing. […] The post Anthropic keeps new AI model private after it finds thousands of external vulnerabilities appeared first on AI News

Compartilhar:

Anthropic’s most capable AI model has already found thousands of AI cybersecurity vulnerabilities across every major operating system and web browser. The company’s response was not to release it, but to quietly hand it to the organisations responsible for keeping the internet running. That model is Claude Mythos Preview, and the initiative is called Project Glasswing . The launch partners include Amazon Web Services, Apple, Broadcom, Cisco, CrowdStrike, Google, JPMorganChase, the Linux Foundation, Microsoft, Nvidia, and Palo Alto Networks. Beyond that core group, Anthropic has extended access to over 40 additional organisations that build or maintain critical software infrastructure. Anthropic is committing up to US$100 million in usage credits for Mythos Preview across the effort, along with US$4 million in direct donations to open-source security organisations. A model that outgrew its own benchmarks Mythos Preview was not specifically trained for cybersecurity work. Anthropic said the capabilities “emerged as a downstream consequence of general improvements in code, reasoning, and autonomy”, and that the same improvements making the model better at patching vulnerabilities also make it better at exploiting them. That last part matters. Mythos Preview has improved to the extent that it mostly saturates existing security benchmarks, forcing Anthropic to shift its focus to novel real-world tasks–specifically, zero-day vulnerabilities. These flaws were previously unknown to the software’s developers. Among the findings: a 27-year-old bug in OpenBSD, an operating system known for its strong security posture. In another case, the model fully autonomously identified and exploited a 17-year-old remote code execution vulnerability in FreeBSD–CVE-2026-4747–that allows an unauthenticated user anywhere on the internet to obtain complete control of a server running NFS. No human was involved in the discovery or exploitation after the initial prompt to find the bug. Nicholas Carlini from Anthropic’s research team described the model’s ability to chain together vulnerabilities: “This model can create exploits out of three, four, or sometimes five vulnerabilities that in sequence give you some kind of very sophisticated end outcome. I’ve found more bugs in the last couple of weeks than I found in the rest of my life combined.” Why is it not being released? “We do not plan to make Claude Mythos Preview generally available due to its cybersecurity capabilities,” Newton Cheng, Frontier Red Team Cyber Lead at Anthropic, said. “Given the rate of AI progress, it will not be long before such capabilities proliferate, potentially beyond actors who are committed to deploying them safely. The fallout–for economies, public safety, and national security–could be severe.” This is not hypothetical. Anthropic had previously disclosed what it described as the first documented case of a cyberattack largely executed by AI–a Chinese state-sponsored group that used AI agents to autonomously infiltrate roughly 30 global targets, with AI handling the majority of tactical operations independently. The company has also privately briefed senior US government officials on Mythos Preview’s full capabilities. The intelligence community is now actively weighing how the model could reshape both offensive and defensive hacking operations. The open-source problem One dimension of Project Glasswing that goes beyond the headline coalition: open-source software. Jim Zemlin, CEO of the Linux Foundation, put it plainly: “In the past, security expertise has been a luxury reserved for organisations with large security teams. Open-source maintainers, whose software underpins much of the world’s critical infrastructure, have historically been left to figure out security on their own.” Anthropic has donated US$2.5 million to Alpha-Omega and OpenSSF through the Linux Foundation, and US$1.5 million to the Apache Software Foundation–giving maintainers of critical open-source codebases access to AI cybersecurity vulnerability scanning at a scale that was previously out of reach. What comes next Anthropic says its eventual goal is to deploy Mythos-class models at scale, but only when new safeguards are in place. The company plans to launch new safeguards with an upcoming Claude Opus model first, allowing it to refine them with a model that does not pose the same level of risk as Mythos Preview. The competitive picture is already shifting around it. When OpenAI released GPT-5.3-Codex in February, the company called it the first model it had classified as high-capability for cybersecurity tasks under its Preparedness Framework. Anthropic’s move with Glasswing signals that the frontier labs see controlled deployment–not open release–as the emerging standard for models at this capability level. Whether that standard holds as these capabilities spread further is, at this point, an open question that no single initiative can answer. See Also: Anthropic’s refusal to arm AI is exactly why the UK wants it Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and is co-located with other leading technology events including the Cyber Security & Cloud Expo . Click here for more information. AI News is powered by TechForge Media . Explore other upcoming enterprise technology events and webinars here . The post Anthropic keeps new AI model private after it finds thousands of external vulnerabilities appeared first on AI News .

Pontos-chave

  • A Anthropic optou por não liberar o Claude Mythos Preview devido a suas capacidades de segurança cibernética.
  • O modelo demonstrou a habilidade de identificar e explorar vulnerabilidades, o que levanta preocupações sobre o uso malicioso da IA.
  • A colaboração com grandes empresas pode inspirar iniciativas de segurança cibernética no Brasil.

Análise editorial

A decisão da Anthropic de não liberar o Claude Mythos Preview, mesmo após a descoberta de milhares de vulnerabilidades de segurança cibernética, levanta questões cruciais sobre a responsabilidade no desenvolvimento de IA. Para o setor de tecnologia brasileiro, isso pode ser um alerta sobre a necessidade de um maior foco em segurança cibernética, especialmente com o crescimento da adoção de IA em diversas aplicações. A colaboração com grandes empresas de tecnologia e organizações de segurança indica que há um movimento em direção a uma abordagem mais colaborativa na mitigação de riscos, o que pode inspirar iniciativas similares no Brasil.

Além disso, a capacidade do modelo de identificar e explorar vulnerabilidades destaca a dualidade da IA: enquanto pode ser uma ferramenta poderosa para a segurança, também pode ser utilizada para fins maliciosos. Isso sugere que o Brasil deve considerar a implementação de diretrizes éticas e regulatórias para o uso de IA, especialmente em áreas sensíveis como segurança cibernética. O desenvolvimento de políticas que incentivem a transparência e a responsabilidade no uso de tecnologias emergentes será fundamental para garantir que o avanço tecnológico não comprometa a segurança nacional.

Por fim, a iniciativa Project Glasswing pode servir como um modelo para futuras colaborações entre empresas de tecnologia e instituições de segurança no Brasil. O investimento significativo da Anthropic em créditos de uso e doações para organizações de segurança de código aberto pode estimular um ecossistema mais robusto e seguro. Observaremos como essa abordagem colaborativa pode influenciar o desenvolvimento de soluções de segurança cibernética no Brasil, especialmente em um momento em que a digitalização está acelerando em todos os setores da economia.

O que esta cobertura entrega

  • Atribuicao clara de fonte com link para a publicacao original.
  • Enquadramento editorial sobre relevancia, impacto e proximos desdobramentos.
  • Revisao de legibilidade, contexto e duplicacao antes da publicacao.

Fonte original:

AI News

Sobre este artigo

Este artigo foi curado e publicado pelo AIDaily como parte da nossa cobertura editorial sobre desenvolvimentos em inteligência artificial. O conteúdo é baseado na fonte original citada abaixo, enriquecido com contexto e análise editorial. Ferramentas automatizadas podem auxiliar tradução e estruturação inicial, mas a decisão de publicar, a revisão factual e o enquadramento de contexto seguem responsabilidade editorial.

Saiba mais sobre nosso processo editorial