ChainStreet
WHERE CODE MEETS CAPITAL
Loading prices…
Powered by CoinGecko
AI

Anthropic Cuts Claude 4.6 Thinking Depth by 67%, Data Shows

Diagnostic session logs indicate Anthropic limits default reasoning capabilities to manage soaring inference costs.

Anthropic Cuts Claude 4.6 Thinking Depth by 67%, Data Shows

Anthropic faces intense backlash after independent session data shows a massive drop in baseline reasoning for the Claude 4.6 model. The AI developer prioritized cost management over performance, pushing a default configuration that generates faster but significantly shallower code edits.

Key Takeaways
  • Anthropic restricted Claude 4.6 reasoning depth by 67% to mitigate rising inference and operational costs for its AI models.
  • Independent data from AMD director Stella Laurenzo shows code reads per edit dropped from 6.6 to exactly 2.0.
  • The unannounced shift fuels "AI shrinkflation" concerns as developers pay premium rates for performance that requires frequent manual retries.
Listen to this article

Quantifying the Cognitive Collapse in Claude 4.6

Stella Laurenzo, Senior Director of AI at AMD, published an independent analysis of 6,852 code sessions earlier this week. The study quantified a 67% collapse in thinking depth by late February 2026. Code reads per edit plummeted from a baseline of 6.6 to exactly 2.0. Analysts treated the metric as a primary proxy for how thoroughly the Claude 4.6 system reviewed context before executing structural changes.

Laurenzo utilized a specialized script to flag shallow reasoning. The diagnostic tool activated 173 times after March 8. It recorded zero firings prior to the second week of that month. Anthropic executed a quiet update on March 3. The firm shifted the default thinking effort from high to medium. Public community complaints regarding missed context and hallucinations surged right after the change.

Anthropic Cuts Claude 4.6 Thinking Depth by 67%, Data Shows
credits: https://github.com/anthropics/claude-code/issues/42796

Unannounced Optimization and Inflated Costs

The company failed to inform users regarding the backend adjustment. Executives provided an explanation only after Hills’ findings circulated across developer forums. Claude Code head Boris Cherny defended the downgrade. Cherny claimed the medium setting offered a practical compromise for typical workloads. The adjustment inflated operational costs for engineering teams. Shallower outputs forced developers to run frequent manual retries to fix broken code.

Anthropic Cuts Claude 4.6 Thinking Depth by 67%, Data Shows
credits: Reddit r/ClaudeAI

Fewer code reads before editing directly translated to poor context understanding. Power users quickly spotted the performance drop during complex tasks. The delayed corporate response fueled industry narratives regarding “AI shrinkflation.” The practice involves providers maintaining premium subscription pricing while throttling compute resources behind the scenes. Engineering teams responded by restricting the model to basic workloads. Many reverted to the older Claude 4.5 Sonnet architecture for core systems.

Advertisement · Press Release

Genuine News Deserves Honest Attention.

High-conviction projects require an intelligent audience. Connect with readers who value sharp reporting.

👉 Submit Your PR

ChainStreet’s Take

Anthropic built a multibillion-dollar brand on safety and transparency. The decision to quietly restrict backend compute shatters the image of a principled laboratory. Developers pay premium rates for reliable automation. A single API call requiring multiple manual prompts to fix sloppy code breaks the financial math of using generative tools. 

Every frontier lab faces intense pressure to manage inference costs. Anthropic caught the heaviest backlash because management executed the downgrade silently. A 67% drop constitutes a massive feature reduction for its users.

CHAIN STREET INTELLIGENCE

Activate Intelligence Layer

Institutional-grade structural analysis for this article.

FAQ

Frequently Asked Questions

01

What is the Claude 4.6 thinking depth reduction?

Anthropic restricted the computational resources allocated to the reasoning processes of Claude 4.6. Data shows the system now performs 67% fewer context reviews before executing code modifications. This adjustment prioritizes faster response times and lower inference costs over deep logical accuracy.
02

Why does this matter for the software engineering industry?

Engineering teams face increased operational expenses because shallower reasoning leads to higher hallucination rates in complex codebases. Developers must now perform multiple manual retries to fix broken structural changes that Claude 4.6 previously handled correctly. This shift raises the real cost of using Anthropic tools for high-stakes production environments.
03

How did Anthropic execute this change?

Anthropic implemented a quiet backend update on March 3, 2026, shifting the default thinking effort from high to medium. Independent analysis by Stella Laurenzo flagged 173 instances of shallow reasoning immediately following the rollout. Executives only acknowledged the change after diagnostic logs began circulating on developer forums.
04

What are the primary critiques of the downgrade?

Critics argue that Anthropic is engaging in "AI shrinkflation" by throttling compute power while maintaining premium subscription pricing. The lack of transparency regarding the March 3 update damaged the reputation of Anthropic as a principled laboratory. Many users feel the move forces them to pay for a product that no longer meets its marketed specifications.
05

How are developers adjusting to the performance drop?

Many engineering teams are reverting to the Claude 4.5 Sonnet architecture to ensure consistent logical output for core systems. Future adoption depends on whether Boris Cherny and the Anthropic team restore high-effort reasoning as an accessible default. Users are currently restricting the model to basic workloads while exploring alternative frontier labs.

You Might Also Like

CHAINSTREET
🛡
Alex Reeve

Alex Reeve is a contributing writer for ChainStreet.io. Her articles provide timely insights and analysis across these interconnected industries, including regulatory updates, market trends, token economics, institutional developments, platform innovations, stablecoins, meme coins, policy shifts, and the latest advancements in AI, applications, tools, models, and their broader implications for technology and markets.

The views and opinions expressed by Alex in this article are her own and do not necessarily reflect the official position of ChainStreet.io, its management, editors, or affiliates. This content is provided for informational and educational purposes only and does not constitute financial, investment, legal, or tax advice. Readers should conduct their own research and consult qualified professionals before making any decisions related to digital assets, cryptocurrencies, or financial matters. ChainStreet.io and its contributors are not responsible for any losses incurred from reliance on this information.