Did Anthropic Secretly Nerf Claude? The Evidence, the Bugs, and What Actually Happened
Over the past month, thousands of developers noticed something wrong with Claude. Responses felt dumber. Token limits ran out in minutes instead of hours. Claude Code started forgetting context mid-session. And Anthropic said nothing for weeks.
Then the accusations started flying.
What Users Are Saying
The complaints started trickling in around early March 2026 and turned into a flood by April.
On GitHub, a developer filed an issue titled “Opus model quality regression” with detailed session logs showing Claude abandoning tasks mid-way through complex engineering work.
On Reddit, a Claude Pro subscriber paying $200 annually reported that “it maxes out every Monday and resets Saturday. Out of 30 days I get to use Claude 12.”
A Max 5x subscriber said they “used up Max 5 in 1 hour of working, before I could work 8 hours.”
One viral analysis titled “Claude Code Drama: 6,852 Sessions Prove Performance Collapse” documented measurable quality drops across thousands of coding sessions.
The word on X, Reddit, and Hacker News was blunt: Anthropic is deliberately dumbing down Claude to save on compute costs.
The Conspiracy Theory
The narrative that formed in developer forums goes like this:
Anthropic voluntarily lowered the intelligence of its model without telling anyone.
They reduced the token limits.
They tried to push Claude Code into the Max plan only ($100/month minimum).
People got angry on X.
Anthropic said it was just a “test” but did not increase the limits.
Tired of the degradation, people started switching to ChatGPT Codex and other alternatives.
Anthropic noticed the churn.
Then conveniently Anthropic announced they had “investigated and found problems.”
Then they reset the token limits.
The implication: Anthropic cut costs, got caught, and reverse-engineered a technical explanation to cover it up.
What Anthropic Says Actually Happened
On April 23, 2026, Anthropic published a detailed post-mortem acknowledging three separate bugs that degraded Claude performance.
Bug 1 Reasoning Effort Downgrade (March 4): Anthropic changed Claude Code default reasoning effort from “high” to “medium” to reduce latency. They later admitted “this was the wrong tradeoff” and reverted it on April 7. That is 34 days of degraded performance.
Bug 2 Cache Clearing Bug (March 26): An optimization meant to clear cached data from idle sessions once ended up clearing it every single turn. This made Claude forget its own context every response. Fixed on April 10.
Bug 3 System Prompt Change (April 16): A system prompt update designed to make Claude less verbose caused a 3% performance drop across both Opus 4.6 and 4.7. Reverted on April 20.
Because each bug hit a different slice of users on a different schedule, the combined effect looked like broad inconsistent degradation.
Is the Explanation Believable?
The technical explanations are plausible. Reducing reasoning effort from high to medium would absolutely make responses feel dumber. A cache bug that clears context every turn would make Claude forget what it was doing. And system prompt changes routinely cause unexpected quality shifts in LLMs.
What makes developers skeptical is the timeline.
Bug 1 was introduced March 4 and not fixed until April 7. That is 34 days.
Bug 2 was introduced March 26 and not fixed until April 10. That is 15 days.
Bug 3 was introduced April 16 and fixed April 20. That is 4 days.
During this entire period Anthropic said nothing publicly. Users were told their complaints could not be reproduced internally. The post-mortem only came after the backlash reached critical mass.
The Compute Cost Problem
Here is the uncomfortable context that feeds the conspiracy theory.
Anthropic subscription plans charge far less than the actual compute cost of tokens consumed. Sometimes by a factor of 10x or more. Every Claude Code session burns thousands of tokens across multiple model calls. At $20/month for Pro, heavy users were getting far more compute than they were paying for.
In February 2026 Anthropic signed a $25 billion deal with Amazon for 5 gigawatts of compute capacity. But that infrastructure takes time to come online. In the meantime every agentic tool is burning inference at unpredictable rates.
The pricing math simply does not work if power users consume $200 worth of compute on a $20 plan.
The Pro Plan Removal
On April 21 right in the middle of the quality degradation controversy Anthropic quietly removed Claude Code from the $20 Pro plan entirely. No announcement. Just a pricing page edit.
Pro users wanting Claude Code now need the Max plan at $100/month minimum. A 5x price jump.
Anthropic called it a “small test on 2% of new signups.” But the support documentation was already updated.
The timing is hard to ignore. In the span of one month Claude Code users experienced degraded quality, reduced token limits, and then a 5x price increase.
What the Community Is Doing
The backlash has pushed developers toward alternatives.
OpenAI launched ChatGPT Codex as a direct competitor during this window. GitHub Copilot and Cursor continue to gain ground. And the local LLM movement got its strongest real-world argument yet.
If a cloud provider can silently degrade your tools or 5x your pricing overnight, running your own models starts looking a lot more appealing.
The Bottom Line
Did Anthropic deliberately nerf Claude to save compute costs? There is no smoking gun. The three bugs they identified are technically plausible and the post-mortem is detailed.
But the 34-day silence while users complained, the convenient timing of the Pro plan removal, the economic pressure of unsustainable pricing, and the pattern of degradation followed by a price increase. That is a lot of coincidences.
Anthropic has since reset token limits for all subscribers and reverted the changes. Whether that is an apology or damage control depends on who you ask.