GitHub just admitted their 10X capacity plan was not enough. They now need 30X. The CTO published an update today that reads like a war report. Two incidents in the last week: a merge queue bug that corrupted branch state across 658 repositories, and a search outage that killed UI functionality for hours. Both trace back to the same root cause. Agentic coding. Since late December 2025, autonomous AI coding agents have been hammering GitHub's infrastructure at a rate nobody planned for. Repository creation, pull requests, API calls, automation, large-repo workloads - all growing exponentially. And here is the part that makes it interesting: a single pull request can touch Git storage, mergeability checks, branch protection, GitHub Actions, search, notifications, permissions, webhooks, APIs, background jobs, caches, and databases. At scale, small inefficiencies compound. Queues deepen. Cache misses become database load. Retries amplify traffic. GitHub Actions is getting hit especially hard. Agentic workflows spawn long-running, parallel CI sessions that dwarf what human developers generate. Copilot code review now consumes GitHub Actions minutes on top of AI credits. The automation layer was not designed for agents running multi-hour autonomous sessions at this volume. The free ride is ending. Starting June 1, Copilot moves to usage-based billing measured in AI credits tied to token consumption. GitHub has already paused new sign-ups for several Copilot tiers. A quick chat question and a multi-hour autonomous coding session used to cost the same amount. That math does not work anymore. Which raises the real question: is the agentic era sustainable on infrastructure built for humans? GitHub is rearchitecting critical systems, isolating services, migrating off legacy frameworks, and pursuing multi-cloud. But the honest read is that the platform is playing catch-up to a usage pattern that showed up faster than anyone modeled. When your 10X plan lasts four months before you need 30X, the planning horizon itself is broken. The agentic era is not a future problem. It is a right-now infrastructure problem. And someone has to pay for it. #GitHub #AgenticAI #DevOps
GitHub's 30X Capacity Plan: Agentic Era Infrastructure Problem
More Relevant Posts
-
GitHub is sitting in a strange spot right now: critical infra for almost every dev team, but struggling with reliability just as AI agents are flooding it with new load. Availability dropping to "one nine" and a steady stream of outages point to infra that was built for humans, not thousands of bots spinning up repos and hammering APIs in the background. At the same time, a tiny startup like Pierre Computer claims to handle repo creation at a scale that looks tailor-made for agents, not people. If GitHub wants to stay the top git platform for AI-native development, it has to treat agent traffic as first-class. That means an AI-native git layer, better scaling of stateful systems like databases and Redis, and a clear North Star around being the backbone for agentic code lifecycles. The current mix of Copilot branding, internal politics, and no CEO naturally pulls attention away from the boring but essential work of hardening the platform. But it is also worth being cautious with the clean narrative. GitHub runs a very different workload from a greenfield product in closed beta, with years of baggage, enterprise constraints, and a massive ecosystem to keep stable. Self-reported numbers from a startup and a rough month of incidents are not enough on their own to declare the incumbent broken or the new model proven. Shutting down Copilot or slicing away half the product surface sounds decisive, yet could throw away real value while the market is still figuring out how devs and agents should work together. The useful takeaway is not that GitHub is doomed or that an AI-only platform will automatically win, but that infrastructure and product strategy now have to be designed around agents and humans coexisting at scale. Getting that tradeoff right - reliability for everyone, while building new, agent-native primitives with a clear focus - will matter a lot more than any single outage or launch over the next few years. https://lnkd.in/dECY42Vt
To view or add a comment, sign in
-
GitHub was designed for humans. AI agents are breaking it. I ran a batch of 40 AI coding agents against a single GitHub repo last week. Within 90 seconds: rate-limited, merge conflicts on every branch, and three token revocations. The architecture assumes a human opens a PR, waits, reviews, merges. Agents don't wait. Cloudflare just shipped a Git platform built for this exact problem. 𝗧𝗛𝗘 𝗕𝗢𝗧𝗧𝗟𝗘𝗡𝗘𝗖𝗞: GitHub's API rate limits and merge queue assume sequential human workflows — agents operate in parallel at machine speed 𝗧𝗛𝗘 𝗦𝗛𝗜𝗙𝗧: Cloudflare's platform treats concurrent writes, branch isolation, and agent-scoped auth as first-class primitives, not afterthoughts 𝗧𝗛𝗘 𝗦𝗜𝗚𝗡𝗔𝗟: Every major cloud provider is building agent-native infra — the tools we built for human developers don't scale to autonomous ones 𝗧𝗛𝗘 𝗤𝗨𝗘𝗦𝗧𝗜𝗢𝗡: How long before your CI/CD pipeline has more agent committers than human ones? If you're running AI coding agents at any scale, the GitHub bottleneck is real. This isn't about replacing GitHub for human workflows — it's about recognizing that agent workflows need purpose-built infrastructure. Anyone else hitting GitHub's walls with agent workloads? Curious what workarounds you've found. Full code + walkthrough → cloudedventures.com #AIAgents #DevOps #CloudEngineering #GitHub #Cloudflare
To view or add a comment, sign in
-
𝗬𝗼𝘂'𝗿𝗲 𝗯𝗮𝗯𝘆𝘀𝗶𝘁𝘁𝗶𝗻𝗴 𝗽𝘂𝗹𝗹 𝗿𝗲𝗾𝘂𝗲𝘀𝘁𝘀. 𝗖𝗜 𝗳𝗮𝗶𝗹𝘀, 𝘆𝗼𝘂 𝗳𝗶𝘅. 𝗥𝗲𝘃𝗶𝗲𝘄𝗲𝗿 𝗰𝗼𝗺𝗺𝗲𝗻𝘁𝘀, 𝘆𝗼𝘂 𝗳𝗶𝘅. 𝗥𝗶𝗻𝘀𝗲, 𝗿𝗲𝗽𝗲𝗮𝘁, 𝗻𝗲𝘃𝗲𝗿 𝘀𝗵𝗶𝗽. Claude Code quietly shipped a feature that collapses that loop. It's called 𝗔𝘂𝘁𝗼-𝗳𝗶𝘅. Once it's on, Claude subscribes to GitHub webhooks for your PR and responds to every CI failure and review comment without you in the room. 𝗧𝗵𝗲 𝗽𝗿𝗼𝗯𝗹𝗲𝗺: A reviewer leaves four comments. You context-switch into each one, push a fix, wait on CI, repeat. 𝗧𝗵𝗲 𝗳𝗶𝘅: Claude reads each comment, makes the clear edits, asks about the ambiguous ones, and replies to the thread under your GitHub account (labeled as the agent, so reviewers aren't confused). 𝗧𝗵𝗲 𝗽𝗿𝗼𝗯𝗹𝗲𝗺: A flaky test fails for the fourteenth time this sprint on something unrelated to your change. 𝗧𝗵𝗲 𝗳𝗶𝘅: Auto-fix sees the check failure, investigates, pushes a fix. If the answer isn't obvious, it pauses and pings you instead of guessing. 𝗧𝗵𝗲 𝗽𝗿𝗼𝗯𝗹𝗲𝗺: The reviewer comes online when you're offline. The PR stalls for a day. 𝗧𝗵𝗲 𝗳𝗶𝘅: The cloud session stays alive while you sleep. Every event fires in real time. You wake up to a merged PR, not a twenty-item TODO. One thing to audit before you flip it on. If your repo uses comment-triggered automation — Atlantis, Terraform Cloud, custom GitHub Actions on issue_comment — Claude's replies can trigger them. Fine for staging. Dangerous for production infra. The GitHub App is required. /𝘄𝗲𝗯-𝘀𝗲𝘁𝘂𝗽 alone won't cut it — install the App on every repo where you want Auto-fix active. Install the App. Open your next PR. Click Auto-fix in the CI bar. The loop gets shorter from there. #ClaudeCode #Anthropic #DevOps #AIEngineering #GitHub #AgenticAI #AIAgents
To view or add a comment, sign in
-
GitHub Actions - The Only Cheatsheet You Need I have been writing GitHub Actions workflows for years. And I still Google the same things every time. What was that expression syntax again? How do I cancel stale runs? Which action do I use for caching? So I put everything in one place. WHAT IS INSIDE 🔔 TRIGGERS — push, pull_request, schedule, workflow_dispatch, workflow_call, repository_dispatch and filters for branches, paths, tags and event types 📦 CONTEXT — every github.* and runner.* variable you actually use, plus steps, needs and job outputs ⚙️ JOBS & STEPS — needs, if conditions, timeout, environment gates, shell overrides, working directory 🔢 MATRIX BUILDS — parallel OS and version runs, include, exclude, fail-fast 🔐 SECRETS & VARIABLES — secrets vs vars, GITHUB_TOKEN, scope levels, passing secrets into reusable workflows ⚡ KEY ACTIONS — checkout, setup-node, cache, upload-artifact, docker build-push, AWS/Azure/GCP OIDC auth 🧮 EXPRESSIONS — contains, startsWith, toJSON, hashFiles, success(), failure(), always() 📤 OUTPUTS & ANNOTATIONS — GITHUB_OUTPUT, GITHUB_ENV, GITHUB_STEP_SUMMARY, masking, notice/warning/error 💻 GH CLI — run, rerun, watch, secrets, cache, all from terminal ONE THING MOST PEOPLE GET WRONG Pinning actions by tag is a supply chain risk. # risky - uses: some-action@v1 # safe - uses: some-action@a1b2c3d4 # full SHA A tag can be moved. A commit SHA cannot. Save this. Bookmark it. Send it to the person on your team who keeps asking you how to set up caching. Drop a comment if there is something missing and I will add it. #GitHubActions #DevOps #CICD #GitHub #DevSecOps #CloudNative #SoftwareEngineering #Automation #PlatformEngineering #100DaysOfCode
To view or add a comment, sign in
-
-
exactly this. agent scale is a whole new thing. this is like the early days of the web, when we started crushing back ends with self-service transactions. agent business is the new ebusiness, and our backends are going to need to be rebuilt and rearchitected to accomodate that change. the next 100x is coming.
GitHub going down is not the story. The reason it went down is. My read: this is a demand signal dressed up as an infrastructure failure. AI agents are reviewing pull requests, writing code, merging changes, and running workflows at a rate no developer platform was designed for. Nobody got a six-month warning that this traffic was coming. It just arrived. This is what the AI transition looks like at the infrastructure layer. A step function. We're seeing the same pattern at Render. This morning, I was in a Slack channel with a company that sells AI agents to other businesses. They keep hitting our API rate limits. Their agents are doing a genuinely unprecedented volume of work. We raise the limits. Then we raise them again. And the message keeps coming back in very direct terms: give us more compute. Infrastructure built for human-paced usage is colliding with AI-paced usage. The next bottleneck is agent growth. Every company running critical systems will face some version of this. The ones that adapt will be fine. The rest will spend the next two years explaining their status page.
To view or add a comment, sign in
-
mapping end to end processes and thinking about what your new pressure points will be what we need to get good at, and quick! tools like value stream mapping can be super helpful here.
GitHub going down is not the story. The reason it went down is. My read: this is a demand signal dressed up as an infrastructure failure. AI agents are reviewing pull requests, writing code, merging changes, and running workflows at a rate no developer platform was designed for. Nobody got a six-month warning that this traffic was coming. It just arrived. This is what the AI transition looks like at the infrastructure layer. A step function. We're seeing the same pattern at Render. This morning, I was in a Slack channel with a company that sells AI agents to other businesses. They keep hitting our API rate limits. Their agents are doing a genuinely unprecedented volume of work. We raise the limits. Then we raise them again. And the message keeps coming back in very direct terms: give us more compute. Infrastructure built for human-paced usage is colliding with AI-paced usage. The next bottleneck is agent growth. Every company running critical systems will face some version of this. The ones that adapt will be fine. The rest will spend the next two years explaining their status page.
To view or add a comment, sign in
-
The backend load explosion with agents reminds me of what happened to banks with the shift to mobile banking. A technology unlocked a behavioral shift (dopamine powered), which increased demand on services by 10-40x factor. My talk track back then was something like this: "Once upon a time, you got into your horse and buggy and brought your paycheck to the local bank to deposit, maybe twice a month. At some point you got direct deposit and the bank built a website, so you checked it there instead. Then you got a phone and for some reason banks are seeing people check their accounts on payday 40 times to see if the check has cleared." True story. And as James said, it required a massive, multi-year rebuild/re-architecture to accomodate. That was just dopamine-fueled curiosity and a device that let us do something on the train or standing in line because we got bored after 7 seconds. This is going to be so much bigger.
GitHub going down is not the story. The reason it went down is. My read: this is a demand signal dressed up as an infrastructure failure. AI agents are reviewing pull requests, writing code, merging changes, and running workflows at a rate no developer platform was designed for. Nobody got a six-month warning that this traffic was coming. It just arrived. This is what the AI transition looks like at the infrastructure layer. A step function. We're seeing the same pattern at Render. This morning, I was in a Slack channel with a company that sells AI agents to other businesses. They keep hitting our API rate limits. Their agents are doing a genuinely unprecedented volume of work. We raise the limits. Then we raise them again. And the message keeps coming back in very direct terms: give us more compute. Infrastructure built for human-paced usage is colliding with AI-paced usage. The next bottleneck is agent growth. Every company running critical systems will face some version of this. The ones that adapt will be fine. The rest will spend the next two years explaining their status page.
To view or add a comment, sign in
-
Great post from Anurag. It's like whacamole. AI agents unlock productivity... which bursts through the damn and floods the valley below (github, render, and soon more!).
GitHub going down is not the story. The reason it went down is. My read: this is a demand signal dressed up as an infrastructure failure. AI agents are reviewing pull requests, writing code, merging changes, and running workflows at a rate no developer platform was designed for. Nobody got a six-month warning that this traffic was coming. It just arrived. This is what the AI transition looks like at the infrastructure layer. A step function. We're seeing the same pattern at Render. This morning, I was in a Slack channel with a company that sells AI agents to other businesses. They keep hitting our API rate limits. Their agents are doing a genuinely unprecedented volume of work. We raise the limits. Then we raise them again. And the message keeps coming back in very direct terms: give us more compute. Infrastructure built for human-paced usage is colliding with AI-paced usage. The next bottleneck is agent growth. Every company running critical systems will face some version of this. The ones that adapt will be fine. The rest will spend the next two years explaining their status page.
To view or add a comment, sign in
-
GitHub Faces Scaling Issues as AI Development Surges It appears that GitHub has its hands full adjusting to the demands of scaling AI workloads. First, the company paused sign-ups for its Copilot subscription tiers in response to a wave of demand from agentic AI projects. Then it shifted to usage-based pricing to, again, better align revenue with the heavy compute demands of AI projects. Now GitHub is confronting still more infrastructure challenges as it deals with the rapid growth in AI-driven software development....
To view or add a comment, sign in
-
🚀 Why did GitHub build their official MCP Server in Go? If you haven’t seen it yet, GitHub just open-sourced their MCP Server — the official Model Context Protocol server that lets AI agents, Copilot, Cursor, Claude Desktop, and other tools directly read repos, manage issues/PRs, analyze code, monitor workflows, and automate like never before. Repo: https://lnkd.in/ezzwZDQr One of the smartest moves? They wrote it in Go. Here’s why Go was the perfect choice (and why it matters for the future of AI tooling): - Insane concurrency with zero drama AI agents fire off dozens of tool calls in parallel. Go’s goroutines + channels make handling streaming MCP sessions, HTTP events, and GitHub API calls feel effortless — without the complexity of threads or async callbacks in other languages. - Production-grade performance & efficiency Low memory footprint, blazing-fast startup, and compiles to a single static binary. Perfect for both the cloud-hosted version (https://lnkd.in/eKKd4fee) and the self-hosted Docker image. No heavy runtimes, no cold starts, just reliable speed. - Simplicity and reliability at scale Go’s standard library already gives you world-class HTTP, JSON, and crypto support. The codebase stays clean and maintainable — exactly what you want when you’re exposing GitHub’s entire platform to millions of AI interactions. - Battle-tested at GitHub They already ship the GitHub CLI and multiple internal services in Go. Reusing the same language, tooling, and operational knowledge just makes sense for a new critical piece of their AI infrastructure. In short: GitHub didn’t pick Go for hype — they picked it because it’s the language that lets them deliver fast, secure, and scalable AI context to developers without compromise. This is a master class in choosing the right tool for the job when building the next generation of developer platforms. 👉 Try it yourself: https://lnkd.in/ezzwZDQr #GitHub #GoLang #Golang #MCP #ModelContextProtocol #AI #Copilot #DevTools #OpenSource #SoftwareEngineering
To view or add a comment, sign in
More from this author
Explore content categories
- Career
- Productivity
- Finance
- Soft Skills & Emotional Intelligence
- Project Management
- Education
- Technology
- Leadership
- Ecommerce
- User Experience
- Recruitment & HR
- Customer Experience
- Real Estate
- Marketing
- Sales
- Retail & Merchandising
- Science
- Supply Chain Management
- Future Of Work
- Consulting
- Writing
- Economics
- Artificial Intelligence
- Employee Experience
- Workplace Trends
- Fundraising
- Networking
- Corporate Social Responsibility
- Negotiation
- Communication
- Engineering
- Hospitality & Tourism
- Business Strategy
- Change Management
- Organizational Culture
- Design
- Innovation
- Event Planning
- Training & Development