Anthropic Reverts A/B Test Removing Claude Code from Pro Subscription Plan

Here's what it means for you.
If you're a subscriber to Anthropic's Pro plan, your access to Claude Code remains intact, but the recent pricing experiment highlights the fragility of AI service offerings.
Why it matters
This test reflects broader trends in AI service pricing and user engagement, impacting how companies structure their offerings amid rising compute demands.
What happened (in 30 seconds)
- On April 21, 2026, Anthropic conducted a limited A/B test removing Claude Code from its $20 Pro plan for about 2% of new subscribers.
- User backlash erupted on platforms like Reddit and X, leading Anthropic to revert the changes within a day.
- Amol Avasare, Anthropic's Head of Growth, clarified that the test was a response to evolving usage patterns and compute constraints.
The context you actually need
- Claude Code is a tool within Anthropic's AI suite that has gained popularity due to its integration with other services, leading to increased resource demands.
- Usage patterns shifted from short interactions to longer, more complex workflows, necessitating a reevaluation of pricing structures.
- The $100+ Max plan was initially designed for heavy chat users but has absorbed features to accommodate high-demand users, reflecting the need for tier restructuring.
What's really happening
On April 21, 2026, Anthropic initiated a pricing experiment that aimed to address the surging compute demands associated with its Claude Code tool. This tool, integral to the Claude AI suite, had seen explosive adoption following its integration with Opus 4 and Cowork, shifting user engagement from brief chat sessions to prolonged, resource-intensive workflows. As a result, the company faced increasing pressure to manage compute costs effectively.
The decision to remove Claude Code from the $20 Pro plan for approximately 2% of new subscribers was a strategic move to test the waters of pricing adjustments amid these evolving usage patterns. However, the immediate backlash from users was swift and vocal, with many expressing frustration over the perceived devaluation of their subscriptions. Developers, in particular, voiced concerns that the removal rendered the Pro plan "useless," leading to demands for refunds and clarifications.
Anthropic's quick retraction of the changes underscores the delicate balance companies must maintain between pricing strategies and user satisfaction. The backlash was not just a reaction to the removal itself but also highlighted broader concerns about the sustainability of AI pricing models in the face of rising compute costs. As users increasingly rely on AI tools for complex tasks, the pressure on companies like Anthropic to provide value while managing operational costs intensifies.
The incident also reflects a growing trend in the AI sector where companies are experimenting with tiered pricing structures to accommodate varying levels of usage. The $100+ Max plan, initially aimed at heavy chat users, has absorbed high-demand features, indicating a shift in how companies are structuring their offerings to meet user needs. This restructuring is essential as the landscape of AI usage continues to evolve, with longer workflows becoming the norm.
In summary, Anthropic's A/B test serves as a case study in the challenges of pricing AI services amid shifting user engagement patterns. The rapid response to user feedback illustrates the importance of maintaining customer trust while navigating the complexities of operational costs and service offerings.
Who feels it first (and how)
- Developers: Frustrated by the removal of key features, they demand refunds and express concerns over the value of their subscriptions.
- AI Enthusiasts: Users who rely on Claude Code for complex workflows feel the immediate impact of the changes.
- Market Analysts: Observers of AI pricing models are keenly interested in how this incident reflects broader trends in the industry.
What to watch next
- Future pricing experiments: Keep an eye on Anthropic's upcoming tests and adjustments, as they may signal shifts in how AI services are priced and structured.
- User engagement metrics: Monitor how changes in service offerings affect user retention and engagement, particularly among developers and heavy users.
- Competitive responses: Watch for how competitors like OpenAI Codex react to Anthropic's pricing strategies and user feedback, which could influence market dynamics.
Anthropic reverted the changes to the Pro plan following user backlash.
Future pricing experiments will continue as companies adapt to changing usage patterns and compute demands.
The long-term impact on user retention and satisfaction remains to be seen.
Insights by A47 Intelligence
Corporate leadership, finance, technology, and market trends.
"Fortune covers financial trends, leadership, and innovation with a pragmatic editorial approach."
— A47 Editor
Anthropic says engineering missteps were behind Claude Code’s monthlong decline after weeks of user backlash
Anthropic has acknowledged that engineering missteps contributed to the monthlong decline of its Claude Code AI platform, which had previously gained significant trust among developers. This decline follows a period of user backlash over performance ...
Daily AI news: models, tools, and policy.
"Independent outlet tracking the fast pace of AI."
— A47 Editor
Anthropic confirms Claude Code problems and promises stricter quality controls
Anthropic has acknowledged significant quality issues with its AI coding assistant, Claude Code, attributing the decline to three specific errors within the system. The company has since implemented fixes and promised stricter quality controls to add...
Biting coverage of AI/ML software and vendors.
"Known for skeptical, incisive reporting on enterprise tech."
— A47 Editor
Anthropic admits it dumbed down Claude when trying to make it smarter
Anthropic has acknowledged that its AI assistant, Claude, experienced a decline in performance due to system changes and bugs, leading to user complaints about lower-quality responses. The company clarified that this was not an intentional effort to ...
Business and tech news excluding paywalled content.
"High-volume business/tech outlet with frequent AI coverage."
— A47 Editor
Anthropic says Claude Code did get worse — but shoots down speculation it 'nerfed' the model
Anthropic has acknowledged that its AI coding assistant, Claude Code, has experienced a decline in performance, identifying three specific issues that contributed to user complaints. The company has refuted claims that it intentionally 'nerfed' the m...
Focuses on transformative tech, AI, gaming, and startup innovation.
"VentureBeat is respected for its in-depth reporting on AI, startups, and disruptive technologies in Silicon Valley and beyond."
— A47 Editor
Mystery solved: Anthropic reveals changes to Claude's harnesses and operating instructions likely caused degradation
Anthropic has acknowledged that changes to Claude's harnesses and operating instructions likely led to a perceived degradation in performance, commonly referred to as 'AI shrinkflation.' Users reported that Claude exhibited reduced reasoning capabili...
Curated tech headlines including AI stories.
"Influential aggregator surfacing the day’s top tech/AI links."
— A47 Editor
Anthropic says it has fixed three causes of recent Claude Code quality issues: reduced default reasoning, a caching bug, and a system prompt to reduce verbosity (Anthropic)
Anthropic has addressed three identified causes of recent quality issues with its AI coding assistant, Claude Code, which included reduced default reasoning, a caching bug, and a system prompt aimed at reducing verbosity. These adjustments are part o...
In-depth coverage of hardware, software, science, and policy.
"Ars Technica provides expert technology news, hardware reviews, and analysis for a technically savvy audience."
— A47 Editor
Anthropic tested removing Claude Code from the Pro plan
Anthropic is currently testing the removal of Claude Code from its Pro plan due to overwhelming demand for its services, prompting the company to explore new methods of rationing access. This decision reflects the challenges faced by Anthropic in man...
In-depth reporting on tech, policy, and science including AI.
"Respected analysis for technically savvy readers, including AI topics."
— A47 Editor
Anthropic tested removing Claude Code from the Pro plan
Anthropic is currently testing the removal of Claude Code from its Pro plan due to overwhelming demand for its services, prompting the company to explore new methods of rationing access. This decision reflects the challenges faced by Anthropic in man...
Daily AI news: models, tools, and policy.
"Independent outlet tracking the fast pace of AI."
— A47 Editor
Anthropic manager hints that Pro and Max plans are outgrown by today's Claude workloads
Anthropic's Head of Growth, Amol Avasare, indicated that the company's Pro and Max subscription plans for its AI assistant Claude are no longer aligned with current user demands, following a brief removal of Claude Code from the Pro plan that was rev...
Daily AI news: models, tools, and policy.
"Independent outlet tracking the fast pace of AI."
— A47 Editor
Unauthorized users breach Anthropic's restricted Mythos AI model
A small group of unauthorized users has breached Anthropic's restricted AI model, Claude Mythos, reportedly accessing it through a private Discord channel since its announcement. This incident raises significant security concerns regarding the model'...