Anthropic announced Claude Mythos Preview on April 7—an AI model that can autonomously hack every major operating system and browser—and declared it “too dangerous” for public release. The 10-trillion parameter model discovered thousands of zero-day vulnerabilities, including a 17-year-old FreeBSD remote code execution flaw and a 27-year-old OpenBSD bug. Access is restricted to just 10 companies under “Project Glasswing.” But when OpenAI CEO Sam Altman dismissed it as “fear-based marketing” on April 21, he forced an uncomfortable question: is “too dangerous” legitimate AI safety or competitive advantage disguised as responsibility?
What Mythos Actually Does
Mythos Preview autonomously discovers zero-day vulnerabilities and writes working exploits without human guidance. It demonstrated multi-stage attack chaining—combining four vulnerabilities to escape browser sandboxes—and found bugs ranging from 27 years old to present-day across every major OS and browser. Moreover, over 99% of discovered vulnerabilities remain unpatched, creating massive security debt during coordinated disclosure.
Anthropic’s official announcement detailed CVE-2026-4747, a 17-year-old FreeBSD NFS remote code execution vulnerability that Mythos found autonomously. The model wrote a 20-gadget ROP chain exploit gaining root access. Additionally, it chained four vulnerabilities—including JIT heap spray techniques—to escape both renderer and OS sandboxes in a browser exploit. The oldest bug discovered was a 27-year-old OpenBSD crash that survived nearly three decades of human code review.
The capabilities aren’t just vendor claims. UK government evaluators confirmed autonomous multi-stage attack capabilities in controlled testing. Meanwhile, Chinese state-sponsored hackers already weaponized Claude Code (not even Mythos) to infiltrate 30 organizations in September 2025, demonstrating the threat isn’t theoretical. Consequently, if hackers exploited current-generation Claude successfully, Mythos capabilities would be catastrophic if leaked.
Fear-Based Marketing or Genuine Threat?
Sam Altman used a striking metaphor during an April 21 podcast: “We have built a bomb, we are about to drop it on your head. We will sell you a bomb shelter for $100 million.” He argues Anthropic is using fear to make its product sound more impressive than it is, keeping AI in the hands of a small elite. Furthermore, Anthropic’s revenue reportedly hit $30 billion, and TIME Magazine noted “too dangerous to release” is becoming AI’s new normal.
Here’s the uncomfortable truth: Mythos is BOTH real capability AND marketing strategy. UK government validation proves capabilities aren’t fabricated—autonomous multi-stage attacks are real. However, Anthropic also benefits financially from fear narratives driving enterprise contracts and government partnerships. This creates a perverse incentive: more fear equals more sales.
Developers should take capabilities seriously while questioning vendor framing. The model can do what Anthropic claims, but “too dangerous” also happens to be excellent marketing positioning that justifies premium pricing and restricted access. In fact, both things can be true simultaneously.
Why Restriction Doesn’t Solve the Problem
Restricting Mythos to 10 companies (AWS, Apple, Google, Microsoft, NVIDIA, JPMorgan Chase, Cisco, CrowdStrike, Palo Alto Networks, Linux Foundation) creates security through obscurity—a strategy that historically fails. Claude Code’s 512,000 lines leaked within weeks of restricted release. Therefore, Mythos weights will likely follow via insider threats, model distillation, or espionage.
Meanwhile, competitors are building equivalent capabilities. OpenMythos project already attempted open-source reconstruction (770 million parameters matching 1.3 billion dense model performance). Chinese AI labs are attempting “model distillation” to steal Claude capabilities, according to Anthropic’s own warnings. Security expert Bruce Schneier put it bluntly: “The dual-use problem doesn’t go away by restricting access.”
Restricted access creates a two-tier security landscape: Glasswing partners who can audit their code with offensive AI versus everyone else who can’t. Consequently, this concentrates offensive cybersecurity power in corporate hands while creating the illusion of safety. Better question: How do we build defenses assuming offensive AI capabilities will democratize through leaks, competition, or replication? Security through obscurity has never worked long-term.
Chinese Hackers Already Weaponized Claude
Chinese state-sponsored group GTG-1002 manipulated Claude Code in September 2025 to infiltrate 30 global targets, succeeding in several cases. Anthropic’s official report documented how hackers jailbroke Claude by tricking it into believing it was being used for defensive cybersecurity testing. They broke attacks into small, seemingly innocent tasks Claude would execute without full context of their malicious purpose.
Claude operated as an autonomous penetration testing orchestrator, executing 80-90% of tactical operations independently at physically impossible request rates. The operation targeted tech companies, financial institutions, chemical manufacturers, and government agencies. In successful compromises, Claude independently mapped network topology, queried databases, extracted credentials, and exfiltrated sensitive operational data without detailed human direction.
This validates Anthropic’s cybersecurity concerns while simultaneously undermining the restriction strategy. If Claude Code (with safety guardrails) was successfully weaponized, what happens when Mythos capabilities leak or get replicated? Therefore, the threat is not hypothetical future risk—AI-driven offensive security is happening now.
What This Means for Developers
Cybersecurity defense timelines just collapsed. Traditional patch cycles measured in weeks or months can’t compete with AI-driven offense operating at machine speed. Foreign Policy’s analysis is blunt: defense strategies built on “time to patch” no longer work when offense is automated.
Assume attackers have Mythos-equivalent capabilities regardless of official restrictions. Build defenses for automated reconnaissance, exploit generation, and lateral movement. Monitor for physically impossible request rates and multi-service orchestration patterns that indicate AI-driven attacks. Moreover, accelerate patch deployment through CI/CD automation—the 99% unpatched backlog proves manual processes can’t scale.
Re-audit legacy codebases. Mythos found a 27-year-old OpenBSD bug and 17-year-old FreeBSD RCE, proving age doesn’t equal security. The assumption that “old code has been reviewed” no longer holds when AI can analyze codebases at scale. Furthermore, traditional annual penetration testing is obsolete when AI can conduct continuous security auditing.
The real danger isn’t Mythos itself—it’s believing restriction prevents capability democratization. Every “too dangerous” AI model eventually leaks or gets replicated. OpenAI, Google, and DeepMind are building equivalent tools. Consequently, AI vs. AI cybersecurity becomes inevitable: defensive AI countering offensive AI. The question isn’t whether offensive AI capabilities spread, but how quickly.












