Published on

I'm on Claude's $20 Plan. Here's How I Made It Feel Like I Have No Limits.

Authors
  • avatar
    Name
    Andrew Blase
    Twitter

The $20 plan isn't the problem. Most people just use it wrong.

I'm an AI startup founder. I use Claude daily — for code, architecture decisions, copy, strategy. For a while, I kept hitting the context wall. Long sessions degrading. Claude forgetting what we built three decisions ago. That creeping sense that the model was working against me, not with me.

Then I started treating token efficiency like a system problem, not a Claude problem. Seven changes later, my output on the same $20 plan is unrecognizable.

Here's exactly what I changed.


Why Most People Burn Through Their Claude Limits Fast

Claude's context window is finite. The longer a conversation runs, the more degraded the model's attention gets. It's not a bug — it's how transformers work. Every token you load is a token fighting for Claude's attention. Old, stale context crowds out the thing you actually need the model to focus on.

Most people open one long chat and live in it all day. By hour three, Claude is hallucinating variable names and forgetting the architecture decision you made 45 minutes ago. They blame the model. The real problem is context hygiene.

Fix the hygiene. Everything else follows.

Developer managing multiple Claude AI chat windows efficiently from a cozy home office

Technique 1: Start a New Chat. Constantly.

This sounds almost insultingly simple. It's the most impactful thing on this list.

Every task is its own chat. Finished debugging that function? Close the chat. Moving to a new feature? New chat. Starting a different codebase? New chat.

A fresh chat gives Claude full, undivided attention. No residual context pulling focus. No accumulated token debt from earlier turns dragging the model toward stale assumptions.

How to implement it: Treat chats like browser tabs — one per task, not one per day. When a task is genuinely done, don't keep piling new requests into the same thread. Open a fresh one. Paste in only what's actually relevant.

The discipline feels slightly awkward at first. Within a week, you won't remember doing it any other way.


Technique 2: Tell Claude to Use Agents on Every Task

Claude can delegate. Most people don't ask it to.

When you're building something non-trivial, ask Claude to break the task into sub-agents instead of trying to handle everything in one linear pass. This does two things: it keeps your main context lean, and it keeps Claude from going down rabbit holes that contaminate the primary thread.

Instead of: "Build me an authentication system with JWT, refresh tokens, rate limiting, and email verification."

Try: "Use agents for this task. Break it into sub-tasks, handle each one separately, and report back with the decisions made."

The model will scope the task, dispatch sub-problems cleanly, and your main context stays narrow and focused. More signal. Less noise.

Why it works: Sub-tasks consume their own context budget. The main thread stays clean. When something breaks, you debug a small sub-context instead of untangling a 200-turn conversation.


Technique 3: Compress at Natural Breakpoints

Every task has natural resting points — a feature ships, a decision gets made, a phase ends. That's when you compress.

Ask Claude to summarize what was decided, what was built, and what state the system is in — then start a new chat with just that summary as context. You've preserved the signal and ditched the noise. The next session starts sharp.

The rule I use: When a task is complete or when context hits roughly 50% full, I trigger compression before moving on. Don't wait until you're at 90% and Claude is struggling to hold the thread together.

How to implement it: At a natural stopping point, send this:

"Before we move on — summarize what we built, key decisions made, current system state, and anything I need to carry into the next session."

Copy that summary. New chat. Paste as the first message. You're back at full attention with zero wasted tokens.


Technique 4: Use Plan Mode Before You Execute

This one saves me the most pain.

Plan Mode is simple: before Claude starts executing anything, you ask it to produce a plan first. You review it. You flag anything that's going in the wrong direction. Then you approve it.

The alternative — just letting Claude run — means it commits to an implementation path, generates hundreds of lines of code, and then you realize the architecture is wrong. Now you're spending tokens unwinding bad work and rebuilding.

How to implement it: Before any non-trivial task, explicitly say:

"Don't write any code yet. Give me a plan for how you'd approach this."

Review the plan. Push back on anything that's off. Then give Claude the green light. You'll save yourself from the worst token-burning spiral there is: redoing something that should have been scoped correctly in the first place.

Developer using plan mode to organize AI agents before executing tasks in Claude

Technique 5: Apply Token-Efficient Rules Across All Your Repos

One of the most practical things I've adopted is Claude Token Efficient Rules — a set of rules you apply globally across your repos that instruct Claude to be structurally less verbose.

What this actually changes: Claude stops over-explaining. It stops restating the task before answering it. It stops padding responses with meta-commentary about what it's about to do. You get tighter, more direct output by default.

How to implement it: Fork or copy the rules from the repo and drop them into your Claude configuration. The difference in response density is immediate. You're getting more code, fewer words — which is almost always what you actually want when you're building.


Technique 6: Use the Serena JetBrains Plugin for Agent-Based Coding Workflows

If you're coding in a JetBrains IDE, the Serena plugin is worth knowing about. It integrates directly with Claude and is built around agent-based workflows — meaning it handles the orchestration layer that most people manage manually or not at all.

Where this helps specifically: instead of pasting large chunks of code into Claude and watching the context explode, Serena manages the context handoff cleanly. It's built for exactly the kind of sub-agent, breakpoint-compression workflow I've described above, but surfaced inside your IDE rather than managed manually in Claude's chat interface.

Best fit for: founders who are doing the bulk of their coding in JetBrains products and want the workflow systematized rather than relying on their own discipline.


Technique 7: Use VoltAgent's Pre-Built Subagent Library

Building custom subagents from scratch takes time. The VoltAgent Awesome Claude Code Subagents repo gives you a library of pre-built subagents you can drop directly into Claude Code.

Instead of reinventing patterns for things like code review agents, test generation agents, or documentation agents, you grab a proven template, modify it to your codebase, and ship.

Why this matters for token efficiency: A well-scoped subagent does one thing cleanly. It doesn't drift. It doesn't pull in context it doesn't need. The pre-built ones in this library are already scoped correctly — you're not starting from a blank prompt that might overreach.


The System, Not the Hacks

Here's what ties all seven together: they're not tricks. They're a system.

  • Fresh chats + compression → Context stays clean
  • Agents + Plan Mode → Work stays scoped before it runs
  • Token-efficient rules + Serena + VoltAgent → Tooling handles the hygiene you'd otherwise forget

Any one of these helps. All seven together means your $20 plan is doing what most people assume requires a Pro account or an API bill.

The model hasn't changed. The workflow has.

Developer closing chat sessions and keeping context clean to optimize Claude token usage

What to Do This Week

Pick the two techniques you're not doing yet. Start there.

If you're running long single-session chats, start fresh. If you're letting Claude execute without a plan, add the plan step. If you're not compressing at breakpoints, add one sentence to your workflow.

The gains compound fast because they're multiplicative — cleaner context means better output, better output means fewer correction cycles, fewer correction cycles means fewer wasted tokens.

Token efficiency isn't a Claude workaround. It's just good workflow discipline for building with AI.

Solo developer with efficient Claude workflow running automated tasks in the background

FAQs

Q: Does starting new chats mean I lose important context?
A: Only if you don't compress before switching. The summary-and-carry approach captures everything you need without dragging along the noise.

Q: Does Plan Mode slow things down?
A: It adds 30 seconds upfront. It saves you from re-doing 20 minutes of wrong-direction work. The math is obvious.

**Q: Are these techniques only useful for Claude's 20plan?A:Nobutthe20 plan?** A: No — but the 20 plan is where the constraints make the discipline feel non-optional. If you're on a higher plan, these habits still make your output faster and better.


Full Stack Data Solutions helps tech workers and solo founders build profitable AI-powered businesses. If you want more workflows like this, follow along on LinkedIn or join the newsletter.