From the source material
1 / 1
Anthropic is positioning Project Glasswing as a defensive coalition for AI-assisted vulnerability discovery. (Image: Anthropic)
Anthropic has announced Project Glasswing, and the useful story is not simply that another AI lab found a scary benchmark to wave around. In Anthropic’s Project Glasswing announcement, the company says it is giving a limited group of major technology, infrastructure, and open-source security partners access to Claude Mythos Preview, an unreleased frontier model that Anthropic says can find and exploit serious software vulnerabilities at a level that changes the cybersecurity calendar. The old schedule was discover, triage, disclose, patch, pray. The new one may be discover at machine scale, then find out whether defenders or attackers built the better factory.
That is why this launch is worth covering as more than corporate coalition bingo. Project Glasswing brings together names including Amazon Web Services, Apple, Broadcom, Cisco, CrowdStrike, Google, JPMorganChase, the Linux Foundation, Microsoft, NVIDIA, and Palo Alto Networks to use Mythos Preview on defensive security work. Anthropic is also extending access to more than 40 additional organizations that build or maintain critical software infrastructure, committing up to $100 million in model usage credits, and donating $4 million to open-source security organizations. Translation: the company is treating vulnerability discovery as an industrial problem, not a heroic-hacker montage.
The alarm bell is Mythos itself. Anthropic says the model has already found thousands of high-severity vulnerabilities, including issues in every major operating system and web browser. In its Frontier Red Team technical write-up on Claude Mythos Preview, Anthropic says the model identified a now-patched 27-year-old OpenBSD bug, a 16-year-old FFmpeg vulnerability, local privilege escalation chains in the Linux kernel, and a FreeBSD NFS remote-code-execution vulnerability that could grant root access. Some of those examples have been patched; many other findings remain under responsible disclosure, which is both reassuring and deeply inconvenient for everyone trying to assess the claim from the outside.
So yes, there is a hype problem here. There is always a hype problem. The phrase “AI era of cybersecurity” practically arrived wearing a conference lanyard. But the underlying shift is real enough to take seriously: advanced coding models are becoming useful not only for writing software, but for reading old, ugly, security-critical codebases with patience humans cannot afford. They can run variations, inspect edge cases, chain clues, and keep doing it after the human researcher would like to go outside. Wonderful. Also: yikes.
The practical point is that vulnerability discovery is only half the story. Finding more bugs faster can make the world safer if the patching, validation, prioritization, disclosure, and deployment machinery scales with it. If that machinery does not scale, it can become a vulnerability confetti cannon pointed at maintainers who already had too much to do. Anthropic says Project Glasswing includes professional human validation before reports go to maintainers, public reporting within 90 days where disclosure allows, and collaboration on recommendations for vulnerability disclosure, software update processes, open-source supply-chain security, secure-by-design practices, regulated-industry standards, triage automation, and patching automation. Good. Those are the boring nouns that matter.
The open-source angle is especially important. Open-source maintainers have spent years being told that their projects are critical infrastructure, usually right before being offered a logo page, a thank-you tweet, and maybe enough funding to buy half a security audit. Project Glasswing’s donations — $2.5 million to Alpha-Omega and OpenSSF through the Linux Foundation, plus $1.5 million to the Apache Software Foundation — are not enough to solve that imbalance. But they point in the correct direction: if AI makes bug finding cheaper, someone has to pay for the human and organizational work of deciding which findings are real, which patches are safe, and how to ship fixes without breaking half the internet.
For security teams, the watch item is not whether they can get Mythos tomorrow. Anthropic says it does not plan to make Claude Mythos Preview generally available, and that is probably wise if its own write-up is directionally accurate. The watch item is how quickly Mythos-class capability leaks into ordinary security tooling, code agents, pentest products, and, eventually, adversarial workflows. Once models can reliably turn old patches into working exploits, or search obscure code paths at scale, “we patched the important stuff eventually” becomes a weaker plan. Patch latency becomes attack surface.
For software teams outside the security department, the uncomfortable takeaway is that AI-assisted coding and AI-assisted breaking are twins. If your organization is accelerating feature work with code agents, it should be accelerating review, dependency hygiene, fuzzing, threat modeling, and update discipline too. More generated code does not automatically mean more insecure code, but faster code without faster assurance is how you build a haunted house and then brag about the square footage.
There is also a governance test for Anthropic. The company is asking the public to accept two ideas at once: first, that Mythos Preview is powerful enough to warrant restricted access; second, that selected partners can use it safely and productively before the rest of the field catches up. That may be the least-bad route. It is still a route built on trust, selective disclosure, and delayed technical details for unpatched bugs. Anthropic will need to show receipts as patches land: what was found, what was fixed, how many reports were false alarms, how maintainers experienced the process, and whether the coalition actually made critical systems harder to attack.
The clickbait version of this story is “AI can hack everything now,” which is not helpful and may not be true. The useful version is narrower and more urgent: AI models appear to be crossing from code assistant into scalable vulnerability researcher, and the defender advantage will depend on whether institutions can absorb that output responsibly. Discovery without repair is noise. Repair without deployment is theater. Deployment without measurement is vibes in a SOC dashboard.
Useful Machines readers should watch Project Glasswing for three things. First, whether the promised 90-day public report gives enough concrete evidence to separate breakthrough from launch-week drama. Second, whether open-source maintainers describe the program as help rather than drive-by bug-report weather. Third, whether the next generation of everyday coding tools ships with stronger security defaults, not just faster autocomplete with better posture.
Project Glasswing is a cyber alarm with a repair plan attached. That makes it more interesting than the alarm by itself. If Anthropic is right about Mythos Preview, the future of software security is not humans versus AI attackers. It is institutions racing to turn AI’s patience, scale, and weird code intuition into patches before someone else turns the same qualities into exploits. Not glamorous. Very necessary. And, for once, the boring part might be the whole ballgame.
In short
Anthropic says Claude Mythos Preview can find and exploit serious software flaws at a new scale. Project Glasswing is its attempt to put that capability in defenders’ hands before attackers get the same advantage.