𝗧𝗵𝗲 𝗔𝗻𝘁𝗵𝗿𝗼𝗽𝗶𝗰 𝘀𝗼𝘂𝗿𝗰𝗲 𝗰𝗼𝗱𝗲 𝗹𝗲𝗮𝗸 is a sobering reminder for those of us building in the LLM space. Your most sophisticated safety guardrails are useless if your CI/CD pipeline fails. Anthropic accidentally bundled a massive debugging file into a Claude Code update. This single mistake exposed 512,000 lines of proprietary code to the public. Within minutes, the leak went viral. A researcher spotted the file, and the news reached over 21 million people before the team could even react. From a technical and implementation perspective, the drama is less interesting than the precedents being set here: - 𝗧𝗵𝗲 𝗜𝗿𝗼𝗻𝘆 𝗼𝗳 𝗨𝗻𝗱𝗲𝗿𝗰𝗼𝘃𝗲𝗿 𝗠𝗼𝗱𝗲: Anthropic built a feature called Undercover Mode specifically to stop Claude from leaking secrets. Yet, a simple packaging error bypassed the very security it was meant to protect. - 𝗧𝗵𝗲 𝗗𝗠𝗖𝗔 𝗟𝗼𝗼𝗽𝗵𝗼𝗹𝗲: When the original repo was taken down, a developer reportedly used AI to rewrite the entire codebase in Python and Rust and named it "𝗖𝗹𝗮𝘄 𝗖𝗼𝗱𝗲". The Python repository hit 50,000 stars almost overnight; it currently sits at 118k stars and 101k forks. This raises a massive question: is an AI-translated rewrite a "new creative work" or a derivative one? - 𝗕𝘂𝗶𝗹𝗱 𝗔𝗿𝘁𝗶𝗳𝗮𝗰𝘁 𝗔𝘂𝗱𝗶𝘁𝗶𝗻𝗴: This wasn't a sophisticated hack, but a production build that accidentally included debugging maps. It serves as a clear signal that when we move at "AI speed", our automated auditing of build packages must be as rigorous as our model alignment. It is easy to look at a giant like Anthropic and wonder how they missed this. However, many teams are one misconfigured config file away from a similar disaster. We must start treating deployment hygiene as a core security function rather than an afterthought. Original thread: https://lnkd.in/gq4aVnyF The Python rewrite repo: https://lnkd.in/gzNCbbgh
Anthropic's 512k Line Code Leak Exposes Claude Secrets
More Relevant Posts
-
🚨 Claude Code Leak — What Happened & Why It Matters Over the past few days, the Anthropic Claude Code incident has sparked serious discussion across the AI and security community. 👉 What happened? A release of Claude Code via npm accidentally included a source map file, which exposed internal TypeScript source code (~500K+ lines). This wasn’t a hack, it was a packaging mistake, but the impact is still significant. 👉 Why this matters: Source maps can reconstruct near-original code Internal architecture + agent workflows became visible Potential for security bypass research (or misuse) A reminder that build pipelines = security surface 👉 Community response: An interesting outcome is this clean-room reimplementation: 🔗 https://lnkd.in/dk9WbNJT The repo recreates parts of the Claude Code agent harness in Python, without using proprietary code, purely based on observed architecture and behavior. 👉 Key takeaway (for engineers & teams): This wasn’t a “breach” — it was a process failure. If you're shipping software: Double-check .map files are excluded from production builds Audit npm/package pipelines Treat CI/CD as part of your security boundary 👉 Sources / coverage: VentureBeat: Claude Code leak breakdown The Register: how the source map exposed code InfoWorld: risks of source maps in production 💭 My take: We spend so much time securing APIs and infra… but incidents like this show that release engineering is just as critical. Curious — how does your team prevent accidental leaks in build pipelines? #AI #Security #DevOps #Anthropic #Claude #SoftwareEngineering #CyberSecurity
To view or add a comment, sign in
-
Handling complexity in long running Python services often feels like juggling fragile glue code, retry loops, watchdogs, and scattered flags. Di Lu’s article, “A supervisor tree library for building predictable and resilient programs,” offers a compelling approach with Runsmith, a Python library inspired by Erlang/OTP supervisor trees that models each unit as a typed worker with an explicit lifecycle. You can read the full breakdown here: https://lnkd.in/dgxjFnpx. What stands out is the shift from brittle process level restarts to fine grained fault isolation and health monitoring that catches stalls and constraint violations, not just crashes. This aligns with challenges I’ve faced building multi component platforms where uptime matters and failure domains must be confined. One caveat is that adopting such a framework requires upfront discipline in designing worker lifecycles and state machines, which can add complexity early on. However, this investment pays dividends when shipping real products that demand maintainability and predictable fault recovery. How have others balanced this upfront design effort against the operational resilience gains in production? #python #softwarearchitecture #systemdesign #reliabilityengineering #productdevelopment #founders #engineering #faulttolerance #opensource #devtools #resilience #longrunningservices
To view or add a comment, sign in
-
🚀 Built & Open-Sourced: A Production-Ready LLM Workflow Engine in Python Over the past few months, I kept running into the same problem: ➡️ LLM apps don’t fail at prompts. They fail at orchestration. What starts simple quickly becomes: – multi-step pipelines – retry logic everywhere – branching flows – zero visibility into execution So I built something focused on that layer. protokol-core 👉 https://lnkd.in/dfA5wZWq 💡 What this actually solves: • Orchestrating multi-step LLM workflows cleanly • Managing retries, failures, and branching logic • Keeping execution fully explicit (no hidden state) • Designing systems that are easy to debug and extend ⚙️ What makes it different: • Minimal, dependency-light core • Composable workflow primitives • Deterministic execution model • First-class support for nested flows / subflows • Clear separation of logic vs execution • Built for real-world systems, not toy demos 🧠 Key idea: LLMs are not the hard part anymore. Workflow + control flow + reliability is. And most frameworks abstract that away… until you actually need control. 🎯 The gap: Most people can build: “prompt → prototype” ✅ Very few can reliably build: “multi-step system → production” That’s where things break. This project is focused entirely on that transition. If you’re working on: • AI platforms • LLM infra • Developer tools • Production AI systems You’ll probably relate to this. Try it. Break it. Extend it. Curious to hear how others are solving this. ⭐ https://lnkd.in/dfA5wZWq #Python #AIEngineering #SystemDesign #LLM #BackendEngineering #BuildInPublic
To view or add a comment, sign in
-
GoScrapy is a high-performance web scraping framework for Go, designed with the familiar architecture of Python's Scrapy. It provides a robust, developer-centric experience for building sophisticated data extraction systems, purposefully crafted for those making the leap from Python to the Go ecosystem. Why GoScrapy? While low-level scraping libraries are powerful, many teams require the high-level architectural framework established by Scrapy. GoScrapy brings this architectural discipline natively to Go, organizing your request callbacks, middlewares, and pipelines into a structured, manageable workflow. Instead of manually orchestrating retries, cookie isolation, or database handoffs, GoScrapy provides the engine that powers your spiders. You focus purely on the extraction logic; the framework manages the high-throughput lifecycle and concurrency in the background.
To view or add a comment, sign in
-
Claude Code's Source Code has been leaked and it's breaking the internet! It's not just an API wrapper of Claude but a tool with multi-level architecture, showing us a very high bar for shipping AI coding tools. So how did this happen? It's Source Maps! Source maps are used for debugging and usually the code that is shipped is minified and compressed to be more abstract. However, source maps map and connect the bundled code back to the original source. NPM accidentally published the source map that effectively shipped the entire source code in human readable format. How to prevent your apps from this mistake? - Audit your NPM before every release using "npm pack --dry-run" - Never include source maps in production packages - Don't overlook .gitignore before pushing changes to production Malicious actors and developers can now better understand the data flow of Claude rather than brute-forcing through prompts injections. Developers can better understand "Claude Code's four-stage context management pipeline and craft payloads designed to survive compaction" (source). The source code was quickly taken down by Anthropic but some were lucky enough to see it before it was ;) Source: https://lnkd.in/g_UYpWfG
To view or add a comment, sign in
-
🚨 DRAMA ALERT 🚨 Anthropic just accidentally open-sourced Claude Code. Twice. Yes, you read that right. The company that called their CLI "secret sauce" and sent hundreds of DMCA takedowns to protect it... published source maps in their npm package. Again. Here's what happened and why it matters for every developer using AI coding tools. 𝗪𝗵𝗮𝘁 𝗮𝗿𝗲 𝘀𝗼𝘂𝗿𝗰𝗲 𝗺𝗮𝗽𝘀? When you ship JavaScript, you typically minify and obfuscate it. Source maps let you reverse that process for debugging. They contain the FULL original source code. Anthropic included these in their npm release. Oops. 𝗧𝗵𝗲 𝗶𝗿𝗼𝗻𝘆 𝗶𝘀 𝗱𝗲𝗹𝗶𝗰𝗶𝗼𝘂𝘀 → Claude Code ranks 39th on Terminal Bench. Dead last among harnesses using Opus. (I love it and it's still my favourite) → The "secret sauce" actually references Open Code's source to match their scrolling behaviour → Competitors copying Anthropic? Nope. Anthropic copying open source projects. 𝗛𝗶𝗱𝗱𝗲𝗻 𝗳𝗲𝗮𝘁𝘂𝗿𝗲𝘀 𝗱𝗶𝘀𝗰𝗼𝘃𝗲𝗿𝗲𝗱 • Dream Mode: Background agents that review past sessions and consolidate memories while you sleep • Coordinator Mode: Spawn multiple parallel workers with shared prompt cache • Kairos: Always-on proactive Claude that monitors your work and can push PRs automatically • Buddy: A Tamagotchi companion that hatches in your terminal (was scheduled for April 1-7) • Undercover Mode: Flag for Anthropic employees to hide that they're using Claude Code in public repos That last one is... interesting. Why hide it? 𝗪𝗵𝗮𝘁 𝘀𝗵𝗼𝘂𝗹𝗱 𝗔𝗻𝘁𝗵𝗿𝗼𝗽𝗶𝗰 𝗱𝗼? The cat's out of the bag. DMCAs won't put it back. 1. Announce a roadmap actually to open source it 2. Let engineers talk about these features publicly 3. Stop the legal theatre 4. Be humans, not lawyers The company that markets itself as the "human" AI lab needs to act like it. Repo fork (while it lasts): https://lnkd.in/gYX8hx9v What's your take? Should all AI coding tools be open source? #ClaudeCode #AIEngineering #OpenSource #SoftwareEngineering #DevTools
To view or add a comment, sign in
-
I’ve been working on an open-source Python library for building AI agents. It’s called Dendrux. The idea is that agent runtimes should handle more than just calling an LLM and tools. In production, you usually need persistence, crash recovery, human approvals, budgets, and guardrails. Dendrux brings it into the runtime. It handles: 1. Tool deny policies and human approval with pause/resume 2. PII redaction at the LLM boundary, so the model sees placeholders while tools receive real values 3. Advisory token budgets with threshold warnings 4. Crash recovery with stale-run sweeping 5. Client-tool bridging for browsers and spreadsheets It’s still early, currently v0.1.0a5, but the foundation is in place. Feedback, issues, and design critiques are welcome. GitHub: https://lnkd.in/gYbhpcdM
To view or add a comment, sign in
-
Your AI code reviews are burning tokens they don't need to. Every time Claude reviews your code, it re-reads the entire codebase. 200 files. 150,000 tokens. For a change that touched 8 files. That's not smart. That's expensive. code-review-graph fixes this. It's an open-source tool that builds a persistent, incremental knowledge graph of your codebase using Tree-sitter and SQLite. Instead of dumping your entire repo into Claude's context, it sends only the changed files plus every file impacted by those changes. The result? 5 to 10x fewer tokens per code review. Before: 200 files scanned, ~150k tokens used. After: 8 changed + 12 impacted files, ~25k tokens used. Here's what makes it practical for real engineering teams: Works natively with Claude Code via MCP (Model Context Protocol). No extra setup, no new workflow to learn. Increments intelligently. After the first build (~10s for 500 files), subsequent updates take under 2 seconds. Only re-parses what changed. Understands blast radius. It traces dependency chains so Claude knows not just what changed, but what else that change could break. Supports 12+ languages out of the box: Python, TypeScript, JavaScript, Go, Rust, Java, C#, Kotlin, Swift, Ruby, PHP, and C/C++. Needs no external database. SQLite is all it takes. The architecture is clean: Tree-sitter parses your code into an AST, a SQLite + NetworkX graph stores the relationships, git diff drives incremental updates, and 8 MCP tools expose everything to Claude Code. Three review workflows ship with it: /code-review-graph:build-graph /code-review-graph:review-delta /code-review-graph:review-pr Whether you're a junior engineer just getting into AI-assisted development or a senior architect thinking about LLM cost optimization at scale, this tool addresses a real problem: context window efficiency. AI code review should be precise. Not brute-force. Check it out: https://lnkd.in/giHvG8pR #AIEngineering #ClaudeCode #LLM #TokenOptimization #CodeReview #OpenSource #DeveloperTools #SoftwareEngineering #MCP #GenAI
To view or add a comment, sign in
-
𝗔𝗻𝘁𝗵𝗿𝗼𝗽𝗶𝗰'𝘀 𝗖𝗹𝗮𝘂𝗱𝗲 𝗖𝗼𝗱𝗲 𝘀𝗼𝘂𝗿𝗰𝗲 𝗰𝗼𝗱𝗲 𝗷𝘂𝘀𝘁 𝗹𝗲𝗮𝗸𝗲𝗱 — 𝗮𝗻𝗱 𝘁𝗵𝗲 𝗱𝗲𝘃 𝗰𝗼𝗺𝗺𝘂𝗻𝗶𝘁𝘆 𝗶𝘀 𝗱𝗶𝘀𝘀𝗲𝗰𝘁𝗶𝗻𝗴 𝗶𝘁 Chaofan Shou (@Fried_rice) spotted the 60MB "cli.js.map" file in the @anthropic-ai/claude-code npm package, which linked to a public Cloudflare R2 bucket containing a "src.zip" of the full unobfuscated codebase. He posted the direct download link on X, alerting the community. https://lnkd.in/dvnqbwwP A clean-room Python rewrite (Claw-Code) went live on GitHub within hours — the fastest repo in history to hit 50K stars (2 hours). Now at 82.2K stars, 81.2K forks, with a Rust rewrite in progress. https://lnkd.in/dPHejfnk 𝗛𝗲𝗿𝗲'𝘀 𝘄𝗵𝗮𝘁 𝘁𝗵𝗲 𝗶𝗻𝘁𝗲𝗿𝗻𝗮𝗹𝘀 𝗿𝗲𝘃𝗲𝗮𝗹 𝗮𝗯𝗼𝘂𝘁 𝗯𝘂𝗶𝗹𝗱𝗶𝗻𝗴 𝗴𝗿𝗲𝗮𝘁 𝗮𝗴𝗲𝗻𝘁𝗶𝗰 𝘀𝘆𝘀𝘁𝗲𝗺𝘀: 1. CLAUDE.md 𝗶𝘀 𝘆𝗼𝘂𝗿 𝗺𝗼𝘀𝘁 𝘂𝗻𝗱𝗲𝗿𝘂𝘁𝗶𝗹𝗶𝘇𝗲𝗱 𝗹𝗲𝘃𝗲𝗿 It is injected on every single turn. You get up to 40,000 characters to encode your architecture, standards, and conventions. Most people barely touch it — that's a mistake. 2. 𝗣𝗮𝗿𝗮𝗹𝗹𝗲𝗹𝗶𝘀𝗺 𝗶𝘀 𝗮 𝗳𝗶𝗿𝘀𝘁-𝗰𝗹𝗮𝘀𝘀 𝗰𝗶𝘁𝗶𝘇𝗲𝗻 Three sub-agent execution models: fork (inherits parent context), teammate (file mailbox), and worktree (isolated git branch). Single-agent workflows are explicitly suboptimal. 3. 𝗣𝗲𝗿𝗺𝗶𝘀𝘀𝗶𝗼𝗻𝘀 𝗮𝗿𝗲 𝗺𝗲𝗮𝗻𝘁 𝘁𝗼 𝗯𝗲 𝗰𝗼𝗻𝗳𝗶𝗴𝘂𝗿𝗲𝗱, 𝗻𝗼𝘁 𝗰𝗹𝗶𝗰𝗸𝗲𝗱 𝘁𝗵𝗿𝗼𝘂𝗴𝗵 Seeing "allow this action?" is a configuration failure. Use settings.json to pre-approve commands. auto mode uses an LLM classifier. --dangerously-skip-permissions is deprecated. 4. 𝗖𝗼𝗻𝘁𝗲𝘅𝘁 𝗰𝗼𝗺𝗽𝗮𝗰𝘁𝗶𝗼𝗻 𝗶𝘀 𝗮𝗻 𝗮𝗿𝘁 Five strategies: micro-compact, context collapse, session memory, full compact, PTL truncation. Use /compact proactively — default 200K tokens, opt into 1M. 5. 𝗛𝗼𝗼𝗸𝘀 𝗮𝘂𝘁𝗼𝗺𝗮𝘁𝗲 𝘄𝗵𝗮𝘁 𝘆𝗼𝘂 𝗿𝗲𝗽𝗲𝗮𝘁 𝗺𝗮𝗻𝘂𝗮𝗹𝗹𝘆 Pre-tool, post-tool, session start/end hooks. Auto-update docs on every commit without prompting. 6. 𝗦𝗲𝘀𝘀𝗶𝗼𝗻𝘀 𝗮𝗿𝗲 𝗽𝗲𝗿𝘀𝗶𝘀𝘁𝗲𝗻𝘁 — 𝘀𝘁𝗼𝗽 𝘀𝘁𝗮𝗿𝘁𝗶𝗻𝗴 𝗳𝗿𝗲𝘀𝗵 Long sessions accumulate structured memory: task specs, file lists, workflow state, errors, learnings. Resume, don't restart. 7. 𝟲𝟲 𝗯𝘂𝗶𝗹𝘁-𝗶𝗻 𝘁𝗼𝗼𝗹𝘀, 𝗽𝗮𝗿𝘁𝗶𝘁𝗶𝗼𝗻𝗲𝗱 𝗯𝘆 𝘀𝗮𝗳𝗲𝘁𝘆 Read-only tools run in parallel; mutating tools run serially. Multiple sub-agents can fan out across your codebase simultaneously. 8. 𝗜𝗻𝘁𝗲𝗿𝗿𝘂𝗽𝘁 𝗲𝗮𝗿𝗹𝘆, 𝗶𝗻𝘁𝗲𝗿𝗿𝘂𝗽𝘁 𝗼𝗳𝘁𝗲𝗻 Streaming means stopping mid-task is cheap. Don't let sunk-cost bias keep a wrong-direction agent running. 𝗖𝗹𝗮𝘄-𝗖𝗼𝗱𝗲 was itself built in a single night using 𝗼𝗵-𝗺𝘆-𝗰𝗼𝗱𝗲𝘅 (OmX) — an agentic harness studying another agentic harness. These architectural ideas will rapidly diffuse into open-source harnesses. That's the beauty of open source. #ClaudeCode #AgenticAI #LLM #AIEngineering #OpenSource #DeveloperTools #MachineLearning
To view or add a comment, sign in
-
More from this author
Explore related topics
Explore content categories
- Career
- Productivity
- Finance
- Soft Skills & Emotional Intelligence
- Project Management
- Education
- Technology
- Leadership
- Ecommerce
- User Experience
- Recruitment & HR
- Customer Experience
- Real Estate
- Marketing
- Sales
- Retail & Merchandising
- Science
- Supply Chain Management
- Future Of Work
- Consulting
- Writing
- Economics
- Artificial Intelligence
- Employee Experience
- Workplace Trends
- Fundraising
- Networking
- Corporate Social Responsibility
- Negotiation
- Communication
- Engineering
- Hospitality & Tourism
- Business Strategy
- Change Management
- Organizational Culture
- Design
- Innovation
- Event Planning
- Training & Development