WIRELOOP
W I R I N G
SIGNAL · SPACE_HDR
March 29, 2026Wireloop BlogNeal Stabell

Anthropic's Mythos leak puts cybersecurity at the center of frontier AI

A misconfigured content system quietly exposed Anthropic's next frontier model, Claude Mythos, and raised fresh questions about how we secure increasingly powerful AI.

Anthropic's Mythos leak puts cybersecurity at the center of frontier AI

Anthropic's Mythos leak puts cybersecurity at the center of frontier AI

In the last week of March 2026, Anthropic's next frontier model, Claude Mythos, went public in the worst possible way: not via a carefully staged launch, but through a misconfigured content system that left thousands of internal assets sitting on the open internet.

What actually happened

On March 26, 2026, security researchers Roy Paz (LayerX Security) and Alexandre Pauwels (University of Cambridge) discovered that Anthropic's content management system (CMS) was exposing a large cache of unpublished material through a publicly accessible data store.

The cache contained close to 3,000 assets linked to Anthropic's blog and research sites—draft posts, PDFs, images, and internal documentation that had never been formally published but were nonetheless reachable via public URLs.

Among those drafts was a long-form blog post describing a new, unreleased model called Claude Mythos, internally codenamed "Capybara."

Fortune reporters reviewed the documents and broke the story in an article published on March 26, 2026, describing Mythos as Anthropic's "most capable" model to date and a "step change" in performance, with particular emphasis on cybersecurity impact.

After Fortune notified the company, Anthropic restricted public access to the data store and attributed the incident to human error in CMS configuration: assets were set to be publicly accessible by default unless manually locked down.

On March 27, 2026, Anthropic's confirmation of the leak and its scope was widely shared in industry posts and newsletters, cementing the timeline: researchers found the cache on the 26th, coverage landed that day, and the company's public acknowledgment followed within roughly 24 hours.

What Mythos is, based on the leak

The leaked draft blog and follow-up reporting sketch a model that goes beyond Anthropic's then-flagship Claude Opus 4.6.

Mythos is described as:

  • A new frontier-class Claude model, built as Anthropic's most capable system so far, and referred to internally as a "step change" in capability.
  • Stronger at long-horizon reasoning and code, with better performance on complex software tasks and multi-step planning than previous Claude generations.
  • Critically, far more capable in offensive security: the draft text reportedly talks about Mythos autonomously discovering vulnerabilities, chaining exploits, and working across operating systems, browsers, and widely deployed software stacks.

That last point is what makes this leak different from "just another model reveal." The leaked materials explicitly frame Mythos as a system whose cybersecurity implications are significant enough to warrant a tightly constrained rollout, including restricted access and specialized governance.

Restricted access and why Anthropic is being cautious

In statements to the press after the leak, Anthropic said Mythos was already in testing with a small group of partners, but not generally available.

The company's own draft blog (as described in reporting) emphasizes that the model's offensive security abilities require "careful, staged deployment" and robust safeguards, not a broad consumer release.

Publicly, Anthropic has framed Mythos as:

  • A model that can dramatically improve defensive security—for example, by helping organizations find latent vulnerabilities in their own systems faster than current tooling.
  • At the same time, a system that could lower the barrier to serious cyber-offense if misused, which is why access is being tightly restricted and channeled through specific cybersecurity initiatives and partnerships.

In early April 2026, Anthropic began to talk more openly about Mythos in the context of a broader cybersecurity program, but still under controlled preview and with explicit references back to the March leak.

That sequence—unintended disclosure, rapid lockdown, then cautious re-introduction on Anthropic's own terms—highlights how central security posture has become to frontier model strategy.

The separate Claude Code source leak

The Mythos story landed just days before a second, unrelated leak put Anthropic's engineering practices under a different kind of microscope.

On the morning of March 31, 2026, security researcher Chaofan Shou noticed that an internal debug artifact in Anthropic's npm distribution appeared to expose a path to a zip archive containing the entire TypeScript source code for Claude Code, the company's popular CLI coding agent.

Within hours, the archive was mirrored into public GitHub repositories; one mirror reportedly accumulated tens of thousands of stars and forks as developers raced to examine the architecture.

Reporting from multiple outlets describes the leaked Claude Code tree as a full production-grade agent stack: around 1,900 files, hundreds of thousands of lines of code, internal feature flags, and orchestration patterns that had not been documented before.

Commentators have called it the first time we have seen such a complete blueprint of a commercial AI coding agent's internals, from tool routing to multi-agent coordination.

Anthropic responded by issuing DMCA takedowns on GitHub. According to TechCrunch's reporting, the initial notice was overly broad and temporarily removed over 8,000 repositories, including legitimate forks of Anthropic's own public Claude Code repo; the company later narrowed the request and GitHub restored the affected forks.

This second incident is separate from the Mythos CMS leak in terms of root cause, but together they tell a coherent story: Anthropic is simultaneously pushing toward more capable, more autonomous systems and grappling with very human, very mundane security failures around the infrastructure that ships and documents them.

Why this matters beyond one company

For customers and builders, the Mythos leak is a reminder that vendor risk for AI is now deeply entangled with cybersecurity risk.

When a single misconfigured CMS can expose the existence and design of a model that might materially change the offensive security landscape, the line between "marketing content" and "high-value technical asset" starts to blur.

At the same time, the Claude Code source leak demonstrates how quickly architectural knowledge propagates once it escapes into the wild: within a day, the patterns Anthropic uses to orchestrate agents, tools, and long-lived coding sessions were being studied, forked, and adapted by the broader ecosystem.

That will likely accelerate innovation around agent frameworks—including open-source projects like OpenClaw and follow-on efforts such as Claw Code, which we will cover in a separate piece focused on agentic tooling.

The through-line is simple: we are entering a phase where the capabilities of frontier models and the security practices around them matter equally. Mythos forces that conversation into the open, earlier than Anthropic planned—but probably right on time for everyone who has to build on, regulate, or defend against the next generation of AI systems.