Interesting Reddit post on r/ClaudeAI.
AI Summary#
This Reddit post from the r/ClaudeAI community discusses a widespread frustration: users on paid plans (Pro and Team) are hitting their usage limits “absurdly fast” when using Claude for serious technical work.
The Core Problem#
The original poster (OP) and many commenters report that even with a disciplined workflow (disabling extra tools/features), their usage quota vanishes within an hour or two of heavy work. The primary “killer” is the cumulative context—Claude reprocesses the entire chat history with every new message, meaning long conversations become exponentially more expensive as they progress.
Key Takeaways & Community Advice#
The community reached a general consensus on how to manage this:
Avoid “Opus” for everything: Users noted that the high-reasoning Opus model burns through the quota roughly twice as fast as Sonnet. The recommendation is to use Sonnet 4.6 for implementation and reserve Opus for high-level planning or final reviews.
Aggressive Context Management: * “Kill your darlings”: Start new chats frequently.
- Handoff Summaries: Before starting a new chat, ask Claude to summarize the current state, decisions made, and next steps. Paste that summary into a fresh session to “reset” the token cost while maintaining continuity.
- Narrow the Scope: Don’t provide your entire codebase if you’re only working on one file. Use
.claudeignoreor explicitly tell the model to focus on specific folders to minimize the tokens “read” during each turn. - Shift to the API: Several power users suggested moving away from the $20/month subscription and using the Claude API instead. This allows for pay-as-you-go pricing, providing more transparency into exactly how many tokens each request costs.
- Treat it like a “Compiler,” not a “Chat”: Instead of small back-and-forth messages (which build up history quickly), try to use “single-shot” prompts that bundle multiple instructions at once.
The Sentiment#
The sentiment is largely negative toward Anthropic’s current tier structure for power users. Many feel the product is “borderline unusable” for complex coding projects because the limits are reached just as the model finally “understands” the full context of the task. Some users even speculated that limits have been tightened recently, though others pointed out this is likely just the reality of the 1M+ token context window being very expensive to maintain.
My Main Takeaways#
/clearreligiously. But first, get summary of prior chat, to seed the next one.- Try
codex? There are claims it’s just as good, if slower; but limits are far more generous! - Fingers crossed a local gemma4 (or better, code-specific) local model will be “good enough”
Additional own observations#
- double check selected model, using bare
/model!- I’ve had instances I did
/model sonnetand it ended up after some days becoming a “custom” model! - (and my observation: these are EXPENSIVE!)
- Always go with “(Recommended)”, unless “you know what you are doing”
- I’ve had instances I did
- don’t forget about
/effort; for mechanical could use “low”?
