For the last three years, $20 a month has been the magic number. Claude Pro, ChatGPT Plus, Gemini Advanced, Copilot Pro, Cursor Pro - all twenty dollars, all clearly priced to anchor against Netflix rather than against enterprise software. That anchor is cracking. The labs are burning cash on inference for power users, the frontier models cost more per token than they did a year ago, and agent tools like Claude Code and Codex are consuming ten to a hundred times the compute a chat session does. Something has to give.

This post pulls together what’s actually happening across the providers, the Claude Code episode that had the developer community briefly losing its mind, and a table of current prices next to what I think the 2026 year-end picture looks like.

The Claude Code Episode

In April, a subset of Anthropic’s $20/month Pro users opened Claude Code and found it locked to Max-tier plans. No announcement. No email. Just a paywall where a working tool had been the week before.

The reaction on Reddit, Hacker News, and X was immediate and loud. Pro subscribers pointed out, reasonably, that Claude Code was one of the main reasons they’d chosen Claude over ChatGPT. Losing it while keeping the same price was a de-facto price hike dressed up as a feature change. Within roughly 24 hours Anthropic clarified that this was a test on a small cohort of users, not a product change, and the affected users had access restored.

The episode is worth paying attention to for three reasons:

  • It’s a signal, not a glitch. Companies don’t A/B test the removal of a headline feature by accident. They run these tests because someone inside wants to know how much churn would follow a repricing. The answer, based on the public reaction, was “a lot.”
  • Claude Code economics are brutal. A single developer running Claude Code for a full working day can easily pull tens of millions of tokens through the system. At Anthropic’s own API prices, that’s $50 - $150 of inference on a $20/month plan. The unit economics only work if most Pro subscribers barely touch it.
  • Max tier is the intended destination. Anthropic already prices Max at $100/month (5x Pro limits) and $200/month (20x Pro limits) specifically for heavy Claude Code users. The obvious strategic move is to keep narrowing Pro’s agent allowance until heavy users self-select into Max.

Anthropic hasn’t announced a removal and, based on the walk-back, isn’t planning one in the near term. But the drift is clear: the $20 tier is increasingly a chat plan with token-metered agent features bolted on, not an all-you-can-eat pass.

Why the $20 Anchor Is Cracking

Three things have changed since ChatGPT Plus launched at $20 in February 2023.

1. Inference got more expensive at the frontier, not cheaper. The low end keeps collapsing - Haiku, GPT-5 mini, Gemini Flash, DeepSeek - but Opus 4.7, GPT-5, and Gemini 2.5 Pro cost more per serious task than GPT-4 did in 2023. Long context, extended reasoning, and multi-turn tool use all multiply the token count. See Token Economics: Why Costs Aren’t Going Down for the longer version.

2. Agents replaced chat as the headline product. Claude Code, Codex CLI, Cursor’s background agents, Gemini in Workspace - these don’t consume tokens like a chat session does. A 30-second chat reply is maybe 2,000 output tokens. An agent running a 10-minute refactor might burn 500,000. Subscription pricing built for chat usage cannot absorb that.

3. The power-user distribution got more extreme. The top 1% of Claude Pro users generate something like 50x the load of the median user. That ratio didn’t exist with ChatGPT Plus in 2023 because there was no way to be a 50x user - you could only type so fast. Agents removed that ceiling.

Put those together and flat $20 all-you-can-eat was always going to end. The question is how the providers unwind it.

Current Prices and Where They’re Going

This is my read on where each main plan lands by the end of 2026. Prices are in USD. “Current” is what’s on the provider’s pricing page today (April 2026). “Projected (end 2026)” is my estimate based on the direction of travel - not an announcement, not a leak, just pattern-matching from behaviour.

Provider Plan Current Projected (end 2026) Likely change
Anthropic Claude Free $0 $0 Tighter message cap, Sonnet only
Anthropic Claude Pro $20/mo $20 - $25/mo Claude Code usage capped harder, not removed
Anthropic Claude Max 5x $100/mo $100 - $120/mo Positioned as the real dev plan
Anthropic Claude Max 20x $200/mo $200 - $250/mo Held steady, usage gently tightened
OpenAI ChatGPT Free $0 $0 GPT-5 mini only, stricter limits
OpenAI ChatGPT Plus $20/mo $25/mo Codex CLI usage metered, Sora capped
OpenAI ChatGPT Pro $200/mo $200 - $300/mo Unlimited reasoning softened to “very high”
OpenAI ChatGPT Team $25/user/mo $30/user/mo Mild uplift
OpenAI ChatGPT Business $30/user/mo $35 - $40/user/mo Enterprise features pushed here
Google Gemini Free $0 $0 Unchanged
Google Google AI Pro $20/mo $25/mo Absorbs more Workspace features
Google Google AI Ultra $250/mo $250 - $300/mo Held, Veo and Deep Research bundled
Microsoft Copilot Pro $20/mo $25/mo Bundled tighter with M365 Personal
Microsoft M365 Copilot $30/user/mo $30 - $40/user/mo Enterprise uplift more likely than consumer
xAI X Premium+ $16/mo $20/mo Grok 4 usage reduced at base tier
xAI SuperGrok $30/mo $35 - $40/mo Modest increase
xAI SuperGrok Heavy $300/mo $300/mo Held, it’s already at the ceiling
Cursor Pro $20/mo $25 - $30/mo Background agents metered separately
Cursor Business $40/user/mo $45 - $50/user/mo Enterprise features expanded
GitHub Copilot Pro $10/mo $10 - $15/mo Agent mode usage capped
GitHub Copilot Pro+ $39/mo $39 - $49/mo Follows Claude Max pricing gravity

A few notes on how to read this:

  • The $20 tier survives, the experience changes. I don’t think anyone pulls the plug on the $20 anchor point this year. It’s too useful as a marketing number. What shifts is what you actually get - fewer agent tool calls, tighter rate limits on the frontier model, more of the headline features routed through higher tiers.
  • The top tier is where the real money is. Pro-at-$200 (OpenAI), Max-at-$200 (Anthropic), Ultra-at-$250 (Google) - these are the tiers the providers actually want heavy users on, and they are all pricing as if AWS and Snowflake are the reference point, not Netflix.
  • Team and Business tiers are getting squeezed upward faster than consumer. Enterprise price elasticity is much lower, so uplifts land there first and land harder.

Where Each Provider Is Probably Headed

Anthropic has been the clearest signal of the repricing. The Max tier was a soft introduction of metered-by-plan pricing for agents. The Claude Code Pro test was a hard introduction that got pulled. The direction is unmistakable - Pro stays at $20, but Claude Code on Pro becomes a limited-allowance feature rather than an open door, and anyone doing real development work is being funneled to Max. See Anthropic’s pricing page for the current state.

OpenAI has the widest pricing ladder in the industry: Free, Plus at $20, Team at $25/user, Business at $30/user, Pro at $200, Enterprise. Codex CLI is the agent pressure point - same economics as Claude Code, same problem. Expect Plus to get a price bump to $25 during 2026 and Codex usage to become explicitly metered on the Plus tier. Pro keeps its “basically unlimited” framing but the rate limits will quietly tighten. OpenAI pricing is worth watching for the first formal uplift.

Google is in the most comfortable position. Gemini Advanced is bundled inside Google One AI Premium, which bundles Drive storage, photo editing, and Workspace AI. The $20 tier isn’t really a standalone AI product - it’s a Workspace upsell - which makes it harder to reprice cleanly. I’d expect a modest move to $25 and a new emphasis on Ultra at $250 for anyone doing real Gemini CLI or Deep Research work.

Microsoft prices Copilot against Microsoft 365, not against OpenAI. Copilot Pro at $20 is a feature attachment, and Microsoft 365 Copilot at $30/user is where most of the enterprise revenue lives. The consumer tier is least likely to move meaningfully. Enterprise has more room and will probably absorb the uplift.

xAI runs Grok through X’s subscription bundles, which insulates it from direct comparison with Claude Pro or ChatGPT Plus. SuperGrok at $30 is the cleanest AI-only tier. SuperGrok Heavy at $300 already does the “power user” job - I don’t expect it to move.

Cursor and GitHub Copilot are in the weirdest spot. They don’t own a model - they resell someone else’s (often several someone elses). When the underlying API pricing shifts, they have to pass it through or eat it. Cursor has already moved to metered pricing on background agents. Copilot Pro at $10 is basically the cheapest viable AI dev tool and will probably stay there because the brand value of “$10 and it just works” is high. The agent-heavy GitHub Copilot Pro+ tier at $39 is where the repricing lands.

What This Means for Individual Users

If you’re a single developer paying out of pocket, the honest answer is that the $20 chat plan is still a fantastic deal and will probably remain one for the rest of 2026, even if it gets 25% more expensive. You are not the subscriber the providers are trying to reprice - the heavy agent user is.

The move that actually changes your bill is adopting an agent workflow. Running Claude Code or Codex eight hours a day will either push you into Claude Max, ChatGPT Pro, or an API account with a meter attached. That’s where the $20 era is genuinely over, and it was never really the $20 era for that kind of usage in the first place.

Some practical advice:

  • If you’re a chat-only user: stay on the $20 plan you already have. Don’t pre-emptively upgrade.
  • If you’re doing daily agent coding: Claude Max 5x or ChatGPT Pro are almost certainly cheaper than the equivalent API usage, and a lot cheaper than Max 20x unless you’re really pushing it.
  • If you’re bouncing between tools: pay-as-you-go API via OpenRouter or similar, with a spending cap, beats any subscription. See GPU Servers vs API Credits for the maths.
  • If privacy or cost is the real constraint: run open-weights locally. Not frontier quality, but cheap, private, and getting closer every month - see Local vs Cloud AI in 2026 and the Mac Studio Local LLM Guide.

The Bigger Picture

The $20/month price point was set when AI was a text-in, text-out product competing with consumer subscriptions. That product still exists and is still priced accordingly. What’s changed is that the flagship experience is now a multi-hour agent session that touches a codebase, a browser, and half a dozen APIs. That product is priced like developer infrastructure, because that’s what it is.

Expect the next 12 months to be messy - quiet test rollouts, walked-back changes, new tiers that didn’t exist last quarter, and a lot of user confusion about what’s actually included in what. Read the pricing page before you renew. Check what’s metered and what’s unlimited. And if you rely on one provider’s agent tool for daily work, have a fallback - the Claude Code Pro episode is a preview, not a one-off.