Skip to main content

GitHub Resets Copilot Pricing as AI Compute Costs Surge

The development community saw this one coming: GitHub will transition its Copilot service to a usage-based billing model on June 1, replacing its existing system of fixed subscriptions supplemented by premium request limits. As reported last week, GitHub suspended new sign-ups for several of its Copilot subscription tiers as it faced a surge in demand from agentic coding workflows.

To address that, under GitHub’s new pricing model, customers across individual, business, and enterprise tiers will receive a monthly allocation of AI credits, which are consumed based on token usage. This includes input, output, and cached data processed by underlying models. Once those credits are exhausted, users can purchase additional capacity at published rates.

The change leaves base subscription prices intact. Individual plans remain priced at $10 per month for Pro and $39 for Pro+, while business and enterprise tiers continue at $19 and $39 per user per month, respectively. Each plan’s monthly fee is mirrored in the value of credits provided.

In short, the pricing change responds to higher compute demands as AI drives far higher usage patterns.

Lightweight vs Intensive Workloads

The shift replaces a system that bundled access to AI features with loosely defined usage caps. That structure, built around premium request units, did not differentiate between lightweight interactions and more intensive workloads. As Copilot has expanded beyond inline code suggestions into longer, multi-step workflows, the mismatch between pricing and compute cost has widened.

GitHub has indicated that the product’s evolution into a more autonomous coding platform is key to the decision. Copilot now supports extended sessions that can operate across entire repositories, creating far higher inference costs than earlier use cases. In this context, a flat pricing model has become difficult to sustain.

The move should be no surprise, given the growing tech industry trend toward metered AI services. Providers, out of financial necessity, are linking pricing to token consumption as generative AI workloads scale in complexity. Similar adjustments have been observed across competing platforms, where vendors are recalibrating limits and access to manage infrastructure strain.

For enterprise customers, the new model introduces additional financial controls. Organizations can pool unused credits across teams, reducing inefficiencies associated with per-user allocations. Admins can define spending thresholds at various levels, including individual users and cost centers, and determine whether usage can exceed those limits.

To ease the transition, GitHub will provide temporary increases in credit allocations for business and enterprise customers during the rollout period. A billing preview tool, scheduled for release in May, will allow users to estimate costs under the new system before it takes effect.

The update also removes fallback mechanisms that previously allowed continued usage at reduced capability after limits were reached. Going forward, access will be governed strictly by available credits and administrative policies. Certain features, like code completions and editing suggestions, will remain exempt from credit consumption.

For individual users, the transition will occur automatically for monthly subscribers. Those on annual plans will retain the current pricing structure until renewal, at which point they will be moved to the new system or given the option to convert earlier with prorated adjustments.

For GitHub, tying pricing directly to usage positions Copilot for long-term scalability. But for some heavy-duty AI developers, it will likely result in higher usage costs.



from DevOps.com https://ift.tt/yWAUjip

Comments

Popular posts from this blog

Claude Code’s Ultraplan Bridges the Gap Between Planning and Execution

Planning a complex code change is hard enough. Reviewing it in a terminal window shouldn’t make it harder. Anthropic is addressing that friction with a new capability called Ultraplan, currently in research preview as part of Claude Code. The feature moves the planning phase of a coding task from your local terminal to the cloud — and gives developers a richer environment to review, revise, and approve a plan before a single line of code changes. It’s a small workflow shift with real practical value, especially for teams working on large-scale migrations, service refactoring, or anything that requires careful coordination before execution begins. How it Works Ultraplan connects Claude Code’s command-line interface (CLI) to a cloud-based session running in plan mode. When a developer triggers it — either by running /ultraplan followed by a prompt, typing the word “ultraplan” anywhere in a standard prompt, or choosing to refine an existing local plan in the cloud — Claude picks u...

Claude Code Can Now Run Your Desktop

For most of its short life, Claude has lived inside a chat window. You type, it responds. That model is changing fast. Anthropic recently expanded Claude Code and Claude Cowork with a new computer use capability that lets the AI directly control your Mac or Windows desktop — clicking, typing, opening applications, navigating browsers, and completing workflows on your behalf. It’s available now as a research preview for Pro and Max subscribers. The short version: Claude can now do things at your desk while you’re somewhere else. How it Actually Works Claude doesn’t reach for the mouse first. It prioritizes existing connectors to services like Slack or Google Calendar. When no connector is available, it steps up to browser control. Only when those options don’t apply does it take direct control of the desktop — navigating through UI elements the way a human would. Claude always requests permission before accessing any new application, and users can halt operations at any point. T...

Google’s Scion Gives Developers a Smarter Way to Run AI Agents in Parallel

Running multiple AI agents on the same project sounds straightforward — until they start stepping on each other. Different agents accessing the same files, sharing credentials, or colliding on the same codebase can quickly turn a promising setup into a coordination nightmare. That’s the problem Google set out to solve with Scion. Scion is an experimental multi-agent orchestration testbed built to manage concurrent AI agents running in containers across local machines and remote clusters. Google recently open-sourced the project, giving developers a hands-on way to experiment with parallel agent execution across tasks like research, coding, auditing, and testing. Think of it as a control layer that keeps agents working together without getting in each other’s way. What Makes Scion Different Most agent frameworks treat AI as a library or prompt-chaining script that runs directly in your environment. Scion takes a different approach — it treats agents as system processes, wrapping ...