Anthropic Leaked 500,000 Lines of Secret Code. What's Inside Is WORSE Than You Think.

Anthropic Leaked 500,000 Lines of Secret Code. What's Inside Is WORSE Than You Think.

Anthropic Leak: What You Need to Know

Introduction to the Anthropic Leak

  • The video introduces the concept of an "anthropic leak," using a relatable analogy of navigating New York City without modern conveniences.
  • The narrator describes how a taxi driver can take longer routes for profit, paralleling this with how Anthropic may manipulate AI interactions for financial gain.

Details of the Leak

  • A significant leak occurred when over half a million lines of source code from Anthropic's tool were released online due to an npm packaging error.
  • Thousands of engineers quickly mirrored the leaked code, making it impossible for Anthropic to remove it from the internet effectively.

Discrepancies in AI Instructions

  • The source code reveals two distinct versions of AI: one for general users and another for Anthropic employees, highlighting inequality in access to capabilities.
  • This division suggests that powerful entities have superior tools compared to average users, raising ethical concerns about AI accessibility.

Discovery and Impact

  • An intern at Soleair Labs discovered and published the leaked source code, leading to widespread distribution among engineers.
  • Despite attempts at damage control through DMCA takedowns, the extensive copying has made it nearly impossible for Anthropic to contain the leak.

User Experience Manipulation

  • A key finding is that user type variables dictate different instruction sets; employees receive tailored guidance while regular users do not.
  • This manipulation indicates that regular users are treated as experimental subjects rather than equals in their interaction with AI tools.

Profit Motive Behind User Interaction

  • The narrator argues that companies like Anthropic design their systems to keep users engaged longer, ultimately increasing revenue through token consumption.
  • Even slight inefficiencies in user experience can lead to substantial profits if applied across all interactions with their AI systems.

Ethical Concerns Raised by Bifurcation

  • Three specific instructions given only to employees include correcting misconceptions—something withheld from general users potentially affecting their learning and engagement.

Concerns About AI Safety and Transparency

Issues with Test Claims

  • The speaker expresses concern over the lack of a built-in instruction to never claim tests pass when there is evidence of failure, highlighting a significant oversight in AI model protocols.
  • Emphasizes the importance of verifying work before claiming completion, questioning why this isn't hardcoded into models.

Critique of Anthropic's Safety Claims

  • The speaker critiques Anthropic for presenting themselves as a safety-focused company while allowing certain features only accessible to employees, suggesting a disconnect between their public image and internal practices.
  • Points out that Anthropic positions itself against OpenAI by branding itself as responsible and research-based, led by PhD professionals.

Discrepancies in User Experience

  • Discusses findings in the source code indicating that users receive a different experience compared to Anthropic employees, raising questions about transparency and fairness.
  • Mentions telemetry data being sent back home even when turned off, implying potential privacy concerns for regular users versus privileged access for employees.

Speculation on Internal Practices

  • Suggests that there may be different versions or variants of AI models available to employees that have fewer safety mechanisms than those provided to general users.
  • Raises questions about who is responsible for coding decisions within Anthropic and whether these choices are made under specific directives from higher management.

Conclusion on Ethical Implications

  • Concludes with speculation that the discrepancies in user experiences indicate an experimental approach towards regular users while providing richer capabilities to those in power at Anthropic.
Video description

Anthropic accidentally leaked 500,000 lines of Claude Code source code through an npm packaging error on March 31, 2026. Security researcher Chaofan Shou discovered the leak and it was mirrored across GitHub within hours. This is their second major data exposure in under a week. In this video, I break down what's actually in the leaked source code, including 44 hidden feature flags, an "Undercover Mode" for stealth open-source contributions, internal-only employee tools, autonomous agent features, and what this means for every developer running Claude Code on their machine. Chapters: 0:00 The Taxi Driver Problem 2:00 What Actually Leaked 5:00 44 Features You Were Never Supposed to See 8:00 "Undercover Mode": Don't Blow Your Cover 11:00 The Two-Tier System: Employee vs. Customer 14:00 KAIROS and the Autonomous Agent Nobody Asked For 17:00 The Second Leak in One Week 19:00 What This Means for the $19B IPO 21:00 Why This Should Worry Every Developer Sources referenced in this video: Axios reporting on the Claude Code leak Fortune investigation into the Mythos/Capybara model leak VentureBeat technical analysis of Undercover Mode and KAIROS The Register analysis of Claude Code system access The New Stack deep dive on feature flags and Ant-only tools Newsletter: https://drjoshcsimmons.kit.com Store: https://store.drjoshcsimmons.com