Anthropic overhauls Claude pricing as users report decline in model performance

Reviewed byNidhi Govil

11 Sources

Share

Anthropic is transitioning Claude enterprise customers from flat-rate subscriptions to usage-based pricing, potentially tripling costs for some users. The shift comes as complaints about the AI model's performance have escalated sharply since February, with developers reporting quality degradation, quota exhaustion, and service outages that have turned a once-favored coding assistant into a source of frustration.

Anthropic Abandons Flat-Rate Enterprise Pricing

Anthropic has quietly restructured its enterprise pricing for Claude, moving away from bundled token allowances toward a usage-based model that charges customers per token consumed

1

. The change affects enterprise customers at contract renewal, eliminating the legacy "Chat-only seats" and "Standard/Premium seats" that previously included flat monthly fees with usage limits

1

.

Source: PYMNTS

Source: PYMNTS

Under the new structure, seat pricing drops from $200 per month to $20 per month, but all token consumption gets billed at standard API rates on top of the base seat fee

1

. According to The Information, this shift could potentially triple costs for some enterprise customers

4

. Adrien Laurent, CEO of IntuitionLabs, an AI consultancy for the pharmaceutical industry, told The Register that many enterprise clients were already spending significantly on overage, with base seat fees representing only 20 percent of total bills while metered API usage accounted for the remaining 80 percent

1

.

Token Consumption Drives Pricing Overhaul

The pricing restructuring reflects broader challenges with AI demand metrics and capacity constraints. Tokens—the basic unit measuring words and characters in both user prompts and AI model responses—have become a distorted metric as companies optimize for volume rather than outcomes

2

. Using Anthropic's latest model rates, one million tokens of input costs $5, while one million tokens of output costs $25

2

.

Agentic AI tools like Claude Code, which execute multi-step workflows and write code, consume thousands of tokens per session compared to simple chatbot interactions

2

. This usage pattern broke the economics of flat-rate pricing. Anthropic's $200-per-month Max plan became a case study in unsustainable subsidies, with heavy users potentially consuming up to $5,000 worth of usage at published API rates while paying just $200 monthly

2

. On April 4, Anthropic cut off third-party agentic tools like OpenClaw that were routing subscriptions through these plans

2

. Boris Cherny, head of Claude Code, explained the subscriptions "weren't built for the usage patterns of these third-party tools"

2

.

Source: The Register

Source: The Register

User Complaints About Claude Performance Surge

As Anthropic adjusts its billing model, user complaints about decline in model performance have escalated dramatically. When The Register asked Claude itself to analyze quality complaints in the Claude Code GitHub repository since January 2026, the AI model concluded that "quality complaints have escalated sharply"

3

. The analysis found April was already tracking 20+ quality issues in 13 days, putting it on pace to exceed March's 18 complaints—itself a 3.5× jump over the January-February baseline

3

.

Stella Laurenzo, a senior director at AMD, posted a viral complaint on GitHub in February stating that Claude Code could no longer be trusted for complex engineering work, with the model appearing to decline in performance including ignoring instructions and providing incorrect "simplest fixes"

4

. Users alleged Anthropic "sneakily turned down how hard claude thinks before editing code" by changing the default from "high" to "medium" effort without adequate disclosure

4

. Boris Cherny responded on X, calling the allegation "false" and stating the change to medium effort resulted from user feedback about Claude using too many tokens, with the modification included in the changelog and communicated via dialog

4

.

Source: VentureBeat

Source: VentureBeat

Caching Changes Compound Usage Quota Issues

Technical changes to Claude's prompt caching system have further strained usage quotas for customers. In March, Anthropic reduced the time to live (TTL) for the Claude Code prompt cache from one hour to five minutes for many requests

5

. User Sean Swanson documented how this "5m TTL is disproportionately punishing for the long-session, high-context use case that defines Claude Code usage"

5

.

While writing to the five-minute cache costs 25 percent more in tokens and the one-hour cache 100 percent more, reading from cache costs around 10 percent of the base price

5

. Jarred Sumner from Anthropic claimed the change made Claude Code cheaper for "one-shot calls where the cached context is used once and not revisited"

5

. However, Swanson reported being a $200 per month subscriber for over six months without hitting quota limits until March, when the "extra burn rate" made the service "unusable"

5

. Pro users paying $20 per month have reported getting as few as two prompts in five hours

5

.

The large one-million-token context window available on paid plans with Claude Opus 4.6 or Sonnet 4.6 models increases costs, especially with cache misses

5

. Boris Cherny acknowledged that "prompt cache misses when using 1M token context window are expensive" and said Anthropic is investigating a 400,000-token context window by default

5

.

Capacity Constraints and the Cone of Uncertainty

Anthropic CEO Dario Amodei has described a "cone of uncertainty" facing AI companies—data centers take one to two years to build, forcing companies to commit billions now for demand they can't yet verify

2

. "If you're off by a couple years, that can be ruinous," Amodei said on the Dwarkesh Patel podcast in February, adding that "some of the other companies have not written down the spreadsheet"

2

.

Anthropic's response has been moving toward per-token billing so revenue reflects actual usage, while competitors like OpenAI have been making AI cheaper and easier to consume at scale

2

. OpenAI's Nick Turley acknowledged that "having an unlimited plan is like having an unlimited electricity plan. It just doesn't make sense"

2

.

Adrien Laurent from IntuitionLabs warned that "it's also very possible that Anthropic is acquiring customers faster than it can scale capacity, and the unit economics simply don't work at the old prices"

1

. He cautioned that Anthropic may lose interest in serving small organizations and individuals if enterprise spending grows large enough, potentially eliminating the flat-rate consumer plans "that got the whole ecosystem here"

1

. The changes arrive as Anthropic approaches a rumored IPO, with demand for Claude AI services outpacing capacity and forcing adjustments to terms of service and rate limits

1

.

Today's Top Stories

TheOutpost.ai

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

Instagram logo
LinkedIn logo
Youtube logo
© 2026 TheOutpost.AI All rights reserved