Claude's leaked source code is absolutely UNHINGED!

Claude's leaked source code is absolutely UNHINGED!

Anthropic's AI Safety and the Source Code Leak

Overview of Anthropic's Commitment to Safety

  • Anthropic is an AI company emphasizing safety, having created a 30-page policy on responsible AI.
  • The company has previously declined Pentagon contracts due to safety concerns.
  • On March 31st, they accidentally released their entire source code online due to a single line missing in a settings file.

Details of the Source Code Leak

  • This incident marked the second occurrence of such a leak within 13 months, raising questions about internal processes.
  • The leaked code included sensitive features like stealth mode for hiding AI authorship in open-source contributions.

Technical Aspects of the Leak

  • Developers typically scramble code before shipping it; however, a source map linking back to the original version was inadvertently included.
  • A previous similar leak occurred on Cloud Code's launch day in February 2025, where developers found another answer key buried in the download.

Consequences and Reactions

  • The recent leak gained significant attention after being highlighted by an intern at a security firm, leading to millions of views on social media.
  • An open bug report indicated that Anthropic’s build tool may have caused this repeated error during the shipping process.

Insights from the Leaked Code

  • Within the leaked source code were over 1,900 files containing hidden features and unreleased models.
  • Notable unreleased features included "Chyros," which automates background tasks while users sleep, and "Ultra plan," which offloads tasks to remote servers.

Hidden Features and Internal Names

  • The code also revealed internal names for unreleased models like Capiara and Numbat, along with references to Opus 4.7—an unannounced model.
  • A Tamagotchi-style virtual pet named Buddy was discovered within Claude Code, featuring various species and RPG stats.

Undercover Mode: A Controversial Feature

  • One surprising feature was "Undercover mode," designed for employees contributing to public projects without revealing their affiliation with Anthropic.
  • This mode instructively strips any identifiable information from commit messages or pull requests made by employees.

Concerns Over Anthropic's Security Practices

Internal Code Names and Leaks

  • The discussion begins with the concern that companies like Anthropic, which use internal code names, may prioritize secrecy over transparency. This raises questions about whether they are hiding AI authorship from the communities they serve.
  • Just five days before a significant source code leak, another incident occurred where a misconfigured website exposed around 3,000 internal files related to an unreleased AI model named Mythos, which was deemed to pose serious cybersecurity risks.

Company Response and Community Skepticism

  • Anthropic attributed the leaks to human error in release packaging rather than a security breach, claiming it was an honest mistake without any personnel consequences. However, this explanation was met with skepticism from the community.
  • The trustworthiness of a company focused on AI safety is questioned when it makes repeated mistakes in software packaging. This inconsistency raises concerns about their capability to manage artificial general intelligence safely.

Defensive Measures and Community Reactions

  • A "poison pill" mechanism is embedded within Claude Code that sends fake features alongside real ones during server communication to deter competitors from copying its functionality.
  • Following the leaks, some community members advocated for open-sourcing the code while others quickly forked it and began creating Python ports. Malicious actors also exploited the situation by uploading fake software packages that contained malware.

Implications of Code Quality and Development Practices

Insights into Claude Code Functionality

  • The video explores how Claude Code operates behind the scenes; when conversations become lengthy, a secondary smaller Claude processes them to summarize key points while discarding irrelevant information.
  • Internally at Anthropic, employees have access to a powerful tool called Tungsten that allows direct control over virtual terminals—capabilities not available in public versions of Claude.

Security Concerns Arising from Summarization Process

  • A critical finding reveals that poisoned instructions can survive through summaries generated by the second Claude. This means attackers could potentially manipulate outcomes without needing direct interaction with the AI.

Naming Conventions as Safety Mechanisms

  • The function names within the code itself serve as warnings against unsafe practices; for instance, one function explicitly instructs developers not to log secrets.
  • Comments left by developers reveal casual attitudes towards coding standards; one engineer noted uncertainty about performance improvements yet proceeded with deployment anyway.

AI Contributions and Quality Control Issues

Automation vs Manual Processes

  • The head of Cloud Code disclosed that all contributions were generated by AI itself—raising alarms about quality control given past errors in deployment processes.

Verification Methods Employed

  • To catch bugs post-deployment, Anthropic uses Claude as a built-in verification agent designed to run tests on its own code but relies on manual checks instead of automated systems—a practice criticized for being prone to oversight.

This structured summary captures key discussions surrounding Anthropic's recent security incidents and development practices while providing timestamps for easy reference back to specific moments in the transcript.

AI Code Writing: A Double-Edged Sword?

The Complexity of AI-Generated Code

  • The AI is programmed to monitor its own performance, attempting to eliminate laziness in code writing. However, it still allows errors, such as a missing settings line, to slip through.
  • A file named "ripple" contains 5,000 lines of code and relies on 219 imports. This complexity categorizes it as a "god object," which is problematic due to its extensive dependencies.
  • Another file, "print ts," exceeds even this size with nearly 5,594 lines. While large files can be acceptable in some contexts (like the Linux kernel), these are not maintained by expert engineers but rather by AI.
  • The design choice for large files stems from efficiency; splitting code into smaller files slows down the AI's processing time and increases costs. Thus, the current structure is optimized for AI functionality rather than human readability.

Implications of Current Practices

  • There are concerns about whether this approach signifies a troubling trend in software development—where code written and organized by AI may lead to unmanageable complexity that goes unnoticed.

Future Developments in AI Software

  • Anthropic appears to be developing advanced tools beyond simple coding assistance, including Chyros (an always-on agent), Ultra Plan (a project planning tool), and Tungsten Tool (exclusive for employees).
  • Features like undercover mode and poison pill strategies indicate an intention to protect proprietary information while hinting at the creation of an autonomous AI workforce. The release of source code raises questions about transparency and trust in cloud-based solutions.
Video description

ALL THE LEAKS: 👉https://www.sabrina.dev/p/claude-code-source-leak-analysis Anthropic, an AI company built on a "safety first" promise, accidentally leaked 512,000 lines of Claude Code's source code to the public internet on March 31st. The crazy part? It all happened because someone forgot a single line in a settings file —and this is the second time they've made this exact mistake in 13 months. In this video, I dig into the leaked Claude Code repository to uncover what Anthropic was hiding. We look at unreleased features like "Kairos" (an always-on background agent), a virtual Tamagotchi-style pet named Buddy, and a highly controversial "Undercover Mode" designed to hide AI authorship when Anthropic employees contribute to open-source projects. We also analyze the "poison pill" decoy trap designed to deter competitors and look closely at the messy reality of half a million lines of code written and managed almost entirely by Claude itself. Is this a harmless human error, or a massive red flag for a company asking us to trust them with AGI? ⏱️ Chapters 00:00 - Anthropic's Safety Promise & The Leak 01:07 - How Claude Code Leaked via Source Maps 02:20 - Making the Same Mistake Twice 03:15 - Hidden Features: Kairos, Ultra Plan & Buddy 05:17 - The Secret "Undercover Mode" 06:43 - Anthropic's Response & Trust Issues 07:25 - The Decoy Poison Pill 08:31 - How Claude Code Actually Works 09:47 - AI-Written Code: God Objects & Strange Comments 13:15 - Is This the Future of AI Software? Titles for the algo - Tragic mistake... Anthropic leaks Claude’s source code - Analyzing Claude Code source code leak - I analyzed the leaked Claude Code (Anthropic's biggest mistake) - The secret features hiding inside Claude Code - Why Anthropic leaked 512,000 lines of their own code SEO Hashtags #ai #claudecode #anthropic #cybersecurity #claude #technology -- I'm Sabrina Ramonov. I'm on a mission to teach 1 million people AI. I'm also the solo founder of Blotato.com. It's a SaaS app for creators/entrepreneurs to go viral faster with AI, just like me (0 to 500k+ in 6 months solo). How I got here… 0: Born in Quezon City, Philippines 🇵🇭 22 yrs old: Graduated UC Berkeley with degrees in Computer Science and Physics... and $250k in student loans. 23 yrs old: Started Qurious, as Founder/CEO, doing natural language processing (AI) in healthcare. 25 yrs old: Closed pilots with billion-dollar pharmaceutical companies, but failed to reach product market fit. 26 yrs old: Pivoted Qurious a lot, then settled on real-time call guidance for sales and customer support, combining streaming speech recognition (AI) and natural language processing (AI). 28 yrs old: Made the Forbes 30 Under 30 list, while living in a living room of 2BR San Francisco apartment. 30 yrs old: Sold 100% of Qurious to Pegasystems (NASDAQ: Pega) in cash and stock deal ($10M+ total). 31 yrs old: Integrated Qurious with Pegasystems Customer Service to launch Pega Voice AI, now used by 1000s of support agents. 32 yrs old: Started angel investing in VC/seed funds, crypto, and passion areas 🍄 while recovering from burnout by hiking, skiing, and reading. Dropped 50 lbs to get back in shape. 33 yrs old: Started making free content to teach 1 million people AI. 34 yrs old: 0 to 500k+ followers in 6 months with $0 budget, 0 team, $0 paid ads, 0 masterminds. Today: Started https://Blotato.com to distribute 100+ pieces of content per week. 1000s of creators & entrepreneurs use Blotato to go viral faster with AI. Get FREE AI prompts, playbooks, and agents 👉 https://sabrina.dev/p/free