Stop wasting 70–80% of your Claude/GPT context window.
If you’re building in crypto, writing Solidity, auditing protocols, or grinding monorepos — you’re probably burning most of your tokens before your model even does useful work.
This repo is a collection of real-world patterns + tools to reduce hidden token waste in AI-assisted development.
Most token waste doesn’t come from prompts.
It comes from invisible layers:
- Claude.md bloat
- history re-reads
- plugin injection
- tool schema overhead
- unnecessary context loading
- logs and terminal noise
- over-generation loops
Result:
👉 73–80% of tokens are spent before actual reasoning begins.
Claude is powerful for large context (200k–1M tokens),
but uses more tokens per prompt (~2× vs GPT in many cases).
So efficiency = everything.
This repo documents + organizes tools that reduce token waste:
- caveman claude — reduces verbose outputs (~75%)
- rtk (rust token killer) — filters terminal output (~60–90%)
- code review graph — focuses only on relevant code (~49× reduction)
- context mode — stores output externally (sqlite), avoids context pollution (~98%)
- claude token optimizer — reduces full-project prompt cost (~90%)
Hidden waste detection
- token optimizer — finds invisible token leakage
- token optimizer mcp — caching + compression (~95%+)
- claude context — vector search across repo (~40% reduction)
- claude token efficient — enforces minimal context usage
- token savior — symbol-based navigation (~97% reduction)
- Large monorepos → code review graph + token savior
- Heavy logs → rtk
- MCP-heavy workflows → context mode
- Fast wins → caveman + token efficient
A 50k-line Solidity monorepo audit:
- before: context exhausted before useful output
- after: ~49× reduction using structured context filtering
Large logs:
- 300k tokens → <10k after filtering
Not to “use AI more”
But to:
- use fewer tokens
- get faster outputs
- reduce API cost
- increase signal-to-noise ratio
Most devs are not limited by AI quality.
They are limited by context inefficiency.
Fix that → everything improves.
MIT