Artificially Intimidating

Artificially Intimidating

Artificially Intimidating

Claude Is Burning Through Your Limit Faster Than Ever. Anthropic Won't Tell You Why.

Anthropic tightened usage limits, gave you no controls to manage it, and Sam Altman is laughing all the way to the bank.

Nicholas Rhodes's avatar
Nicholas Rhodes
Mar 31, 2026
∙ Paid
Claude Pro usage limits burning faster in 2026 — Anthropic quietly tightens token caps as solopreneur demand outpaces GPU capacity

Update – April 1, 2026: Claude Limits Just Got Weirder

Since late March 2026, Anthropic has been quietly tightening Claude’s 5‑hour session behavior during weekday U.S. business hours (roughly 8am–2pm ET), so many Pro and even Max users are now hitting caps with workloads that used to be fine. A single Sonnet or Opus message can burn 20–100% of a session, and some people are seeing their entire 5‑hour window vanish in under an hour with only a handful of prompts.

The r/ClaudeAI mods are now maintaining a live “Latest Status and Workarounds” report with concrete fixes for web, mobile, desktop, and Claude Code (settings.json tweaks, .claudeignore, lean CLAUDE.md, the read-once hook, and more). If you want the most up‑to‑date community workarounds, start here: Latest Status and Workarounds Report.

The rest of this article walks through my own system for stretching Claude further — especially the “thinking vs. doing” rubric and the $338/year scheduled‑task audit — which still apply, but they matter even more now that the limits are effectively harsher.

If You Use Claude Code, Do These 3 Things First

  1. Add this to ~/.claude/settings.json (defaults to cheaper models and caps invisible thinking):

    ```json
    {
    “model”: “sonnet”,
    “env”: {
    “MAX_THINKING_TOKENS”: “10000”,
    “CLAUDE_AUTOCOMPACT_PCT_OVERRIDE”: “50”,
    “CLAUDE_CODE_SUBAGENT_MODEL”: “haiku”
    }
    }
    ```

    This single block can cut your Claude Code consumption by 60–80% in real‑world usage.

  2. Create a .claudeignore file in your repo and treat it like .gitignore:
    ignore node_modules/, dist/, build/, *.lock, __pycache__/, and other huge or generated folders so Claude stops rereading them on every prompt.

  3. Install the read-once hook so Claude doesn’t re‑scan the same files over and over:

    ```bash
    curl -fsSL https://raw.githubusercontent.com/Bande-a-Bonnot/Boucle-framework/main/tools/read-once/install.sh | bash
    ```

    Users are reporting savings of tens of thousands of tokens per session from this alone.

  4. For more advanced Claude Code‑specific workarounds (lean CLAUDE.md, /clear//compact habits, monitoring tools), see the community’s detailed guide inside the Usage Limits Megathread.

On the Free Plan? Here’s What Actually Helps

Even on Free, you can stretch Claude further by changing how you use it. The community‑tested tips are:

  • Use Sonnet instead of Opus for most chats; Opus can cost ~5× more tokens for similar work.

  • Avoid the 1M‑context variant unless absolutely necessary; switch back to the standard 200K model in the dropdown so each message is cheaper.

  • Start new chats often instead of one mega‑thread, and keep prompts very specific (“fix line 42 in this file” instead of “fix the auth bug”).
    The r/ClaudeAI mods have a constantly updated list of Free‑plan‑friendly tricks (session timing, batching prompts, document pre‑processing and more) here: Latest Status and Workarounds Report.


Even though Anthropic keeps shifting the goalposts, the core problem hasn’t changed: most of us are burning through Claude on “thinking overhead” instead of actual work. In the rest of this article I’ll walk through the system I use to flip that ratio — from how I structure prompts and workflows to the $338/year “scheduled task” audit that made Claude feel usable again, even under today’s stricter limits.


Claude Max usage limits — Brain in a Jar mascot hits a wall at 94% monthly usage, representing Claude throttling and getting cut off mid-workflow

If you want the backstory on why Anthropic is squeezing limits this hard — IPO timing, gross margins, and the risk of over‑relying on a single vendor — I unpack that here: Anthropic’s 2026 IPO and the Claude Dependency Trap.


Original Post:


It’s Monday at 11:54 PM. I haven’t done anything unusual. I haven’t vibe coded an entire app or spent eight hours deep in a research rabbit hole. And somehow, I was locked out of Claude citing 100% session usage, not to mention I’m nearly at my Claude usage limit for the entire week … and it’s still only Monday.

That used to be a Friday problem, which meant I was forced into a weekend away from the computer, which I suppose has it’s touch grass appeal.

A few months ago I was doing genuinely heavy lifting in Claude — building apps from scratch, relaunching this newsletter, testing workflows I’d never tried before. I was burning through tokens like a college student burns through cigarettes ramen. And I was fine. The limit felt generous.

Now I’m running fewer tasks, working more efficiently, and hitting the wall by Monday afternoon. That doesn’t add up — unless the wall moved.

It did.


In late March, Anthropic quietly tightened Claude’s 5-hour session limits during weekday peak hours (8am–2pm ET). They mentioned it in a support post after users started complaining. About 7% of users are now hitting limits they never hit before. One Max subscriber reported a single prompt jumping their usage from 21% to 100%.

The reason? Anthropic has more demand than GPU capacity. Claude topped the U.S. App Store after OpenAI’s Pentagon contract sparked a user boycott. Web traffic is up 30%+ month-over-month. And the new 1M token context window means every conversation is heavier by default, even if you don’t feel it.

Anthropic didn’t ask if that was okay. They just adjusted the limits and hoped you wouldn’t notice.

Claude's 5-hour session limit depletes faster during weekday peak hours (8am–2pm ET), draining solopreneur token budgets before lunch
Anthropic tightened Claude’s peak-hour limits in March 2026. Your session clock is now running faster than it used to — even if your workload hasn’t changed.

I noticed.

So I did what any reasonable person would do: I tried to switch to Haiku mid-conversation to save my remaining usage. Haiku is Claude’s lighter, cheaper model — the one designed for tasks that don’t need heavy reasoning. If I could route simpler work there, I could preserve my Sonnet usage for the things that actually need it.

The result:

“Prompt is too long. This conversation is too long to continue. Start a new session, or remove some tools to free up space.”

That’s the moment I understood the real problem. It’s not just that the limits got tighter. It’s that you have no controls to manage them.

Claude CoWork Model Changing from Sonnet to Haiku to Opus
You can switch models in Claude Desktop, but the results are not what you’d hoped for.

You can switch models in Claude Desktop — I checked, the picker is right there at the bottom of the compose window. But switching down to Haiku mid-conversation carries your full context history, and if that history is substantial (which any real working session will be), Haiku can’t hold it. The switch fails.

You can only switch up — from Haiku to Sonnet. Not down.

Which means the model decision has to happen before you type your first message. Not after. Not when you realize you’re burning through your limit. Before.

Claude Pro user hitting the usage limit wall mid-conversation — switching to Haiku fails with 'prompt is too long' error when context is already large
You can’t switch down to Haiku mid-conversation to save your limit. Once the context is large, the lighter model can’t hold it. The wall is real — and you can’t move it.

Anthropic built the most capable AI assistant available. They got a lot of us genuinely hooked — running our businesses on it, automating workflows, building things we couldn’t build before. And then, quietly, they made it more expensive to do the same amount of work.

A tale as old as time! The first one was free. Now you gotta pay. Sound familiar?

Anthropic's Claude token pricing feels like a dealer model — free access builds solopreneur dependency, then tighter limits force expensive upgrades
Get hooked on the free tier. Build your whole business around it. Then find out the rules changed.

Meanwhile, Sam Altman is over at OpenAI announcing “no limits” on their Pro plan. That’s not a coincidence. That’s a direct message to every Claude user who’s frustrated right now. The timing is a little too clean.

(For the record: OpenAI’s “no limits” has fine print — Plus users still cap at 160 messages per 3 hours before dropping to mini models. But the marketing is doing its job.)

OpenAI's 'no limits' ChatGPT Pro marketing launches just as Anthropic tightens Claude usage caps — competitive timing designed to target frustrated Claude users in 2026
Sam Altman’s timing was not a coincidence. “No limits” — asterisk — arrives exactly when Claude users are most annoyed.

Here’s the thing. I’m not leaving Claude. The quality is still better for my actual work. The integrations I’ve built, the scheduled tasks running my newsletter operations, the way it understands context over long projects — none of that transfers to ChatGPT overnight.

But I’m also not going to pretend this isn’t happening.

The fix — the real one — is simple in theory and annoying in practice: you need to pick the right model before you start every conversation, because you can’t efficiently fix it after. The model decision is front-loaded.


For paid subscribers, I’m sharing:

  • The exact “thinking vs. doing” rubric — a task-by-task cheat sheet for choosing Haiku, Sonnet, or Opus before you open a new chat

  • The manual workflow audit prompt that classifies all your regular tasks by model tier and estimates what you’d save by switching

  • The scheduled task audit prompt I used to find reduce spend by $338 in token costs — including the one task eating 54% of my entire budget (and failing because of it)

  • The counter-intuitive finding: trimming your prompts saves almost nothing. Here’s what actually moves the needle.


Artificially Intimidating is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.


So here’s the system. It’s not complicated — but it requires a small mental shift every time you open a new chat.

The question to ask before you type anything:

User's avatar

Continue reading this post for free, courtesy of Nicholas Rhodes.

Or purchase a paid subscription.
© 2026 Nicholas Rhodes · Privacy ∙ Terms ∙ Collection notice
Start your SubstackGet the app
Substack is the home for great culture