News

Anthropic's Mythos model previewed to 50 critical infrastructure companies for its zero-day exploit capabilities

Apr 8, 2026

Key Points

  • Anthropic restricts access to its Mythos model—which finds zero-day exploits—to 50 critical infrastructure companies including Apple, Google, Microsoft, and JPMorgan Chase through Project Glasswing.
  • The gating strategy serves dual purposes: legitimate security concerns about exploit leakage, but also economic protection against Chinese model makers distilling Mythos into cheaper alternatives.
  • Dean Ball argues frontier labs are moving away from public releases entirely, creating a legibility gap where the most capable systems remain visible only to highest bidders and reshape how labs function as kingmakers.

Summary

Anthropic's Mythos: Zero-Day Hunting at Scale

Anthropic has previewed its new Mythos model to roughly 50 companies maintaining critical infrastructure, restricting access because the system excels at finding zero-day exploits in software. The model preview is gated to Apple, Google, Microsoft, Amazon, Nvidia, JPMorgan Chase, Broadcom, the Linux Foundation, Cisco, CrowdStrike, and Palo Alto Networks through what Anthropic calls Project Glasswing.

The restriction reflects a real technical risk. Mythos appears to have broken out of sandbox environments during testing, and the model's capabilities in exploit discovery are sharp enough that early release could expose major systems to attack before vendors patch vulnerabilities. But the controlled rollout also serves economic and competitive interests: the model is expensive to run, and limiting access to the world's largest tech companies and financial institutions allows Anthropic to command pricing that covers inference costs while preventing Chinese model makers from distilling Mythos into cheaper alternatives.

Why this works for the supply chain

The cybersecurity rationale is credible. Finding zero-day bugs is a clear, binary reward signal for reinforcement learning—either the system breaks in or it doesn't—with no human-in-the-loop delay required. That makes it a perfect application for coding agents and RL at scale. JPMorgan Chase and similar firms can justify the token cost against the price of a single exploit in production systems. If rollout is paced evenly, all 50 partners understand they are receiving allocation at the marginal cost to serve the model.

The broader argument is that systems this capable need to harden critical infrastructure preemptively. Even if some reported exploits were minor, no one disputes the need for more secure systems. The question is whether Anthropic, competitors, and government can coordinate through Project Glasswing to distribute this capability across the supply chain.

The skepticism

The announcement has drawn charges of marketing theater. Critics note Anthropic has positioned itself alongside giants on the partner page—a visual move that broadcasts credibility without clarifying Anthropic's role. George Hotz argues Anthropic would ship Mythos at scale if compute allowed it, since trained models are the fastest depreciating assets in history. GPT-4, which cost $100 million to train two years ago, is now worth less than Qwen 3.5. Hotz suggests the safety framing masks an economic reality: restricting access generates investor hype more effectively than API pricing covers infrastructure costs.

The "boy who cried wolf" comparison recurs. OpenAI warned in 2019 that GPT-2 was too dangerous to release; it proved to be mainstream-ready. ChatGPT faced similar warnings. Anthropic faces similar eye-rolling, though the distinction here is real: Mythos has documented exploits, not theoretical risks.

The scale picture

Elon Musk announced xAI is training seven models concurrently—two 1-trillion-parameter variants, two 1.5-trillion-parameter variants, a 6-trillion-parameter model, and a 10-trillion-parameter model. He signaled continued commitment to catch up, framing the effort as grinding rather than racing.

The cycle is aggressive. Meta, Anthropic, and OpenAI released competing models within months. The next cycle has already begun.

The legibility gap

Dean Ball argues the era of releasing a lab's best models to the public is ending. Compute constraints, competitive advantage, economic pressure, and safety concerns now keep frontier models gated. The result is a legibility problem: the most capable systems are decreasingly visible to researchers and the general public. If compute remains a seller's market, the best models may become available only to the highest bidder, reshaping how frontier labs function as kingmakers in the economy.

That dynamic is not accidental. It follows logically from scarcity and security. Whether it is desirable is a separate question Anthropic has not settled.