People with a Max plan often ask this: running Claude Code in the terminal all day barely makes a dent in their quota. Switch to Cowork for a few hours on a small project and you’re already hitting the limit.
Same Max plan. Same Claude. What’s going on?
The Short Answer
Cowork burns tokens dramatically faster than the Claude Code CLI. One Cowork session doing complex file operations can consume the same quota as dozens of regular chat messages. Max 5x’s “225+ messages” translates to roughly 10–20 real operations in Cowork.
There are three layers to why.
Layer One: Cowork’s Hidden Token Overhead
Cowork runs inside a sandboxed VM. Behind every action, things you don’t see:
- Screenshots and image processing: Vision tokens are expensive — far more than plain text
- Multiple AI inference calls: A single “task” can trigger 5 to 10 API calls
- Full file reads into context: No fine-grained control over what gets loaded, unlike the CLI
- Compounding history: Every step carries the full history of all previous steps — context only grows
You think you’re doing one thing. Underneath, it’s running a dozen API calls.
Layer Two: Claude Code CLI Is Inherently Token-Efficient
The CLI has advantages Cowork doesn’t:
Prompt caching. System prompts, CLAUDE.md, and tool definitions — the stuff that repeats across turns — gets cached. You don’t pay for it again each session.
Precise context control. You decide which files to read, which skills to load. A well-structured CLAUDE.md with on-demand loading is dramatically leaner than dumping everything into context at once.
No vision processing. Pure text interaction — no screenshots, no image recognition. That entire category of cost simply doesn’t exist.
Layer Three: Slowness Creates a Feedback Loop
Cowork is slow. A task that Claude Code finishes in 5 minutes can take 40 minutes in Cowork.
Slow isn’t just frustrating — slow means more accumulated context, which means more tokens per step. The longer a task drags on, the more history every subsequent step has to carry. It compounds.
What the Community Has Documented
This isn’t just anecdotal. GitHub issues:
- #16856: Users reporting token consumption 4x higher than before
- #23318: Multiple people confirming abnormal usage, suspecting billing changes
- #33120: Cowork-specific rate limit issues
Zvi Mowshowitz wrote a dedicated piece comparing the two, reaching the same conclusion: the gap comes down to context control. On Threads, developers said “Burns limits way faster than Claude Code” and “5min task in CC → 40min in Cowork.”
Extra Observation: Why Opus Feels Smarter in the CLI
This is subjective, but there’s a reasonable explanation.
CLI context is cleaner and more focused. The model can spend its compute on the actual problem instead of processing sandbox environment noise. Same model, better input, better output.
When to Use Which
Claude Code CLI: Development work, long sessions, tasks where you want precise control. Highest token efficiency. Built for daily heavy use.
Cowork: GUI operations, browser interactions, or when you’re not comfortable with the terminal. The tradeoff is significantly higher quota consumption.
If you have a Max plan and your main work is coding, you rarely need to touch Cowork.