← All Posts

There you have it. Anthropic — the company building the most popular AI coding assistant on the p…

April 1, 2026 · 0 likes · 0 comments
AI Cybersecurity
There you have it. Anthropic — the company building the most popular AI coding assistant on the planet, running at $2.5 BILLION in annual revenue — shipped their entire Claude Code source code to the public because someone forgot to exclude a source map from their npm package.

512,000 lines of code. 1,900 TypeScript files. The full agentic harness. The slash commands. All of it. Downloaded. Forked 41,500+ times before Anthropic could even respond. 21 million views on X.

The internet doesn't forget.

And here's the thing — this isn't the first time. The first leak happened in February 2025. Same root cause: source maps left in a production package. They had 13 months to add a CI check that strips map files before publish.

They didn't.

But this isn't even the most interesting part.

The leak exposed 44 feature flags for fully-built but unshipped features. Not vaporware. Compiled code sitting behind flags, ready to ship:

— Background agents running 24/7 with GitHub webhooks
— One Claude orchestrating multiple worker Claudes
— Cron scheduling
— Voice command mode
— Full browser control via Playwright
— Agents that sleep and self-resume
— Persistent memory across sessions

It's all done. The "roadmap" is a release schedule.

Now here's what nobody in mainstream coverage is talking about:

That leaked harness — the software layer that governs Claude's behavior, validates permissions, and implements the guardrails — is now a white-box attack surface. A malicious AI agent pointed at those 512,000 lines tonight can systematically map every trust boundary in hours. Every edge case in the permission validation logic. Every one of those unshipped feature flags. Then generate proof-of-concept exploits automatically.

That's not science fiction.

Traditional 0-day discovery required skilled humans reading code for weeks. Not anymore. The threat model has changed. You point an adversarial AI at the code and let it work overnight.

And Claude Code is an agent with real OS-level permissions inside enterprise environments, connected to codebases, credentials, and internal systems. A 0-day in the harness isn't just "Claude is jailbroken." It's potential lateral movement into whatever environment Claude Code is operating in.

Anthropic says this "was not a security breach." Technically true about the packaging incident. But the downstream implications of that code being in the wild ARE a breach-level event. Second incident in under a week — five days ago: 3,000 internal files leaked including details on their next model "Mythos," described as a "step change in capabilities" with unprecedented cybersecurity risks.

At some point, this stops being bad luck.

If you're deploying AI agents in enterprise or government — this is your reminder. Security doesn't just apply to the model. It applies to your entire engineering practice. The pipeline IS the product.

I'll be breaking this down live on April 9th — join us: https://lnkd.in/eG2jWvPf

What are your thoughts?
View original on LinkedIn →