Original Reddit post

I’ve found few projects compression tool that does the job but not sure if they’re completemary or can be used in parallel Most claim the same RTK: “CLI proxy that reduces LLM token consumption by 60-90% on common dev commands.” https://github.com/rtk-ai/rtk Caveman: “Claude Code skill that cuts 65% of tokens by talking like caveman” https://github.com/JuliusBrussee/caveman Repomix: “packs your entire repository into a single, AI-friendly file” https://github.com/yamadashy/repomix Context-mode: “Context window optimization for AI coding agents. Sandboxes tool output, 98% reduction.” https://github.com/mksglu/context-mode Codebase-Memory-mcp: “Indexes codebases into a persistent knowledge graph — average repo in milliseconds. 155 languages, sub-ms queries, 99% fewer tokens” https://github.com/DeusData/codebase-memory-mcp Headroom: " https://github.com/chopratejas/headroom " https://github.com/chopratejas/headroom Lean CTX: “Reduce more by 60–95% (up to 99% on cached reads) Shell Hook + MCP Server” https://github.com/yvgude/lean-ctx Distill: “Distill large CLI outputs into small answers for LLMs and save tokens!” https://github.com/samuelfaj/distill Sigmap: “97% token reduction for AI coding sessions — zero deps, 21 languages, MCP server” https://github.com/manojmallick/sigmap submitted by /u/zakblacki

Originally posted by u/zakblacki on r/ClaudeCode