Copilot’s new usage-based pricing: what 27x model multipliers mean for engineering teams
GitHub Copilot is shifting from flat-rate generosity to usage-based reality. A widely shared Reddit post argues that GitHub’s revised “multiplier” table is the warning shot before full usage billing kicks in. The headline change: Opus 4.6 requests are now priced at a 27x rate, and Sonnet 4.6 at 9x, based on how quickly they drain your monthly pool of premium requests.
If your team has been defaulting to the biggest models for everything, expect your forecasts to wobble. For UK organisations, this isn’t just a developer convenience story – it’s a budgeting, governance and data protection story.
What changed: model multipliers and “premium request” pools explained
Copilot plans include a pool of premium requests each month. Different models drain that pool at different rates, based on a “multiplier”. According to the Reddit post, GitHub has significantly increased the multipliers for Anthropic’s latest Claude family models used in Copilot.
| Model | Previous multiplier | Current multiplier | Practical impact |
|---|---|---|---|
| Claude Opus 4.6 | 3x | 27x | Drains premium pool ~9x faster than before |
| Claude Sonnet 4.6 | 1x | 9x | Drains premium pool 9x faster than before |
Jargon check:
- Tokens – the units models process and bill for (think chunks of text). More tokens = more cost.
- Context window – the total text the model can consider at once. Long context often means higher token usage.
- Agentic workflows – multi-step, tool-using, partially autonomous tasks, not just a single prompt and reply. These are typically far more compute-intensive.
“The 27x multiplier is closer to honest pricing.”
Why this is happening: subsidy unwind and compute constraints
The post’s core claim is simple: the flat-rate era was subsidised. Compute costs, especially for long-context and agentic features, have outpaced what users were paying under one-size-fits-all plans. Anthropic is described as “compute-constrained”, with advanced workflows reportedly consuming 10–100x more tokens per user than basic chat.
Infrastructure at this scale is slow to expand. The post suggests it takes 18–24 months to build capacity to match demand, during which time providers (Microsoft and Anthropic, in this case) were “absorbing the gap”. The new multipliers are a move towards cost-reflective pricing.
“The free lunch is over. Adjust your defaults before June 1!”
Usage-based billing: what UK engineering leaders should expect
According to the post, GitHub will move to full usage-based billing on 1 June. The multipliers are a signal; the billing shift is the budget event. Many organisations have provisioned Copilot widely as a benefit, without:
- Visibility into which models developers are invoking.
- Per-user usage dashboards.
- Policies for model selection or cost caps.
If your team has been running Opus “on everything” – code review, boilerplate, quick completions – costs will spike under usage billing. Expect finance to ask why AI spend is off-forecast. The Reddit author predicts some teams could see a 15x variance.
Balanced view: frontier models can be worth it. For gnarly refactors, complex codebase reasoning or deep reviews, Opus-level capability can produce material productivity and quality gains. The point is selectivity. Use the right model for the right job.
How to prepare: cost controls and model governance before June
Set sensible model defaults
- Default to a mid-tier model for routine tasks (e.g. one-line completions, simple Q&A) and reserve frontier models for complex work.
- Encourage prompts that constrain scope and context length to cut token use.
Implement visibility and guardrails
- If your tooling supports it, enable organisation-level logs that include user, model, and request counts.
- Add soft limits or alerts by user/team. Fallback to cheaper models when thresholds are approached.
- Tag AI usage to cost centres or projects so spikes are traceable.
Review long-context and agentic features
- Long-context sessions and automated multi-step agents are the cost outliers. Use them intentionally, not as the default.
- Break large tasks into smaller, cheaper steps where possible.
Build a minimal usage dashboard
- Even a lightweight tracker helps. If you need a quick approach, you can adapt this guide to log usage events to Google Sheets: How to connect ChatGPT and Google Sheets.
UK-specific considerations: compliance, procurement and data
- GDPR/DPA 2018 – treat Copilot providers as data processors. Review what code and metadata are sent, retention policies, and data residency options (not disclosed in the Reddit post).
- Security – check how secrets are handled in prompts and context. Audit prompts for inadvertent PII or confidential data leakage.
- Procurement – if you buy via enterprise agreements or frameworks, confirm how usage charges will be invoiced and capped. Align billing cycles to your budget controls.
- Education – brief engineers on model selection and cost-aware prompting. A 10-minute cheat sheet can prevent expensive defaults.
What this signals for the wider AI market
The post argues that every major provider is “unwinding” flat-rate plans, with structures designed so heavy agentic usage reflects its true cost. Expect more granular pricing, more controls, and more pressure to prove ROI. For teams that built workflows assuming near-unlimited frontier access, this is the moment to refactor.
Two outcomes are likely positive: clearer economics should drive better tooling for metering and governance, and it should nudge teams to choose appropriately-sized models, which often improves latency and developer experience.
Action checklist for the next two weeks
- Audit your current Copilot model defaults and long-context features.
- Enable whatever usage logs, alerts and limits your plan supports.
- Create a model policy: default, when to escalate, and when to fall back.
- Tag usage to projects/cost centres; prepare a simple dashboard.
- Brief teams on cost-aware prompting and context control.
Sources and further reading
- Original discussion: Copilot just 9x’d Sonnet and 27x’d Opus and teams have no idea
- Vendor documentation and pricing pages (for current terms and controls):
Note: all figures and timelines cited are from the Reddit post. If you rely on them for budgeting, verify against official GitHub communications and your enterprise contract.