Cycled through Claude Code’s effort dropdown last week, hunting for config glitches. There it was: Max, lurking above High like a secret boss level nobody mentioned.
My team’s been cranking .NET libraries—GPU transpilers, WebRTC stacks, ML inference—on what we thought was peak Claude power. $200 a month for Max 20x, the crown jewel of Anthropic’s individual plans. High used to mean full-throttle reasoning, exhaustive logic traces. Now? It’s capped. Efficient. Prone to statistical guesses that crater at runtime.
Anthropic slipped this in late March 2026. No email. No CLI banner. Just a quiet relabel, demoting your max to middle-tier without a whisper. For blog posts or toy React apps, who cares? But transpile C# IL to WebGPU shaders? Debug WebRTC race conditions? High spits out code that looks right—until it buffer-aliases your GPU into crashes or leaks memory like a sieve.
Why Did High Suddenly Suck?
Here’s the table from the front lines. My team audited post-switch:
| Metric | On High (Before) | On Max (After) |
|---|---|---|
| Test suite | 471 passing | 498 passing |
| WebRTC P2P | Broken (offer collision) | Working (offers-with-announce) |
| Wire extensions | Untested | 27 new tests, all passing |
| BEP 46 ECDSA | Gaps | 10 tests with real crypto verification |
| Production bugs | 0 (shipped blind) | 24 found in one sweep |
Those 24 bugs? Buffer aliasing nuking WebGPU. Unreturned GPU rentals leaking everywhere. Reflection dispose torching shared accelerators. WebRTC signaling? Total rewrite—Claude on High pattern-matched generics, ignoring my reference repos. Max? It reasoned through the architecture, nailed it.
“The Max setting caught all of these. High missed them because it was ‘being efficient’ — which is another way of saying it was taking shortcuts.”
That’s straight from Todd Tanner, the .NET dev who blew this open on his blog. He’s shipping SpawnDev libs, open-source beasts for local GPU compute.
Market dynamics scream warning. Anthropic’s burning cash on frontier models—Claude 3.5 Sonnet, Opus—while inference costs balloon. Compute’s the moat, but it’s also the noose. Add Max above High? Boom—instant cost savings on the old top tier. Subscribers foot the bill, debugging AI hallucinations that High now greenlights.
You’re not just paying $200. You’re double-dipping: subscription plus engineer hours chasing ghosts. My bet? This erodes fast. Look at AWS EC2 bursting— they relabel instance types, nudge you up tiers quietly. SaaS 101: Sell premium, then redefine it downward when margins pinch. Anthropic’s no different. Intelligence-as-a-Service means your “max” shrinks with their GPU tab.
And those in-CLI polls? “How well is it doing? 1-3, dismiss.” Data goldmine. Every dismiss trains them on minimum viable smarts. Sneaky.
Is Anthropic’s Claude Pricing Sustainable for Devs?
Short answer: Nope, not for heavy engineering. Premium plans promise message limits and depth. But when High flips from “full reasoning” to “budget cap,” it’s bait-and-switch. Tanner clocked 46% quota at mid-week on High—headroom galore. Lower effort wasn’t quota-saving; it was Anthropic’s compute shield.
Zoom out. AI dev tools market’s a $10B frenzy—GitHub Copilot at $10/pop, Cursor scaling wild, local inference like Ollama exploding. Anthropic’s at 5-10% share, per SimilarWeb traffic, but retention? This stunt risks it. Devs notice runtime fails. We fork to open weights or self-host.
Unique angle: Echoes Netflix’s 2011 Qwikster flop. They splintered DVD/streaming tiers mid-subscription, no notice. Churn spiked 20%. Anthropic’s playing tier-creep roulette—add Max, watch High users grumble or bolt. Prediction: By Q4 2026, 15% premium churn if notifications stay MIA. Sovereign stacks like Tanner’s SpawnDev.ILGPU (WebGPU/CUDA/OpenCL on your iron) fill the void. No corp throttling your think cycles.
Check your CLI now. High? Flip to Max. Audit shipped code—re-run on true max. Defaults betray.
But here’s the rub. Anthropic’s not evil; they’re scaling. Claude’s reasoning depth costs exponentially—Max chews 5-10x the tokens, per leaked benchmarks. Still, silence kills trust. Email your tier leads. Demand changelogs.
Why Does This Matter for AI Coding Tools?
Devs, you’re the canary. Casual users tap Claude for emails, fine on High. But production code? Max or bust. Broader: Validates local-first push. Why rent brains when OpenAI’s o1-preview hints at chain-of-thought blowups, and Anthropic pulls this? SpawnDev’s my hedge—GPU ML inference on your box, no quotas, no downgrades.
Market shift incoming. Gartner pegs edge AI at 30% CAGR through 2030. Anthropic et al. fight inference wars; winners own the stack. Losers? Tier-trapped subs debugging AI shortcuts.
Fix your settings. Build sovereign. Or keep paying for quiet-quitting AIs.
🧬 Related Insights
- Read more: Kubernetes 1.35 Sneaks Safer CSI Tokens Past the Logs — Without Breaking Your Setup
- Read more: The Smartest Apps Hide Their Power: Less UI, More Magic
Frequently Asked Questions
What is Claude’s Max effort setting? Max delivers exhaustive reasoning in Claude Code, above the now-capped High tier. Essential for complex code like GPU transpilers or WebRTC—catches bugs High misses.
Did Anthropic notify users about the new Max tier? No. Added late March 2026, no emails, banners, or CLI alerts. Existing High users got downgraded silently.
Should I switch to Max in Claude Code? Yes, if doing prod engineering. Audit recent High-generated code—expect bugs. Defaults won’t save you.