Original Reddit post

I’d been bouncing between Claude, ChatGPT, and Gemini for a while, but the current Claude rate limit situation finally forced me to clean it up properly. The main problem wasn’t switching tools. It was losing context every time I switched. A few people in my last post said something that ended up being the real unlock: memory and skills are just files So I stopped thinking about this like a chat problem and started thinking about it like a system design problem. What I built: • central LLM wiki for shared context • shared skills folder used across models • one main agent file + one main md file controlling the overall setup • repo-specific agent files + CLAUDE.md files for local context So every project has its own context, but everything still points back to one main source of truth. And it’s working way better than I expected. My wiki is getting built out super nicely, all my models are using it much more consistently, and the whole setup feels actually scalable now. This is the first cross-model setup I’ve had that doesn’t feel like chaos. If you’re dealing with rate-limit bouncing too, I’d seriously recommend thinking in terms of: shared wiki + shared skills + global control files + repo-specific overrides That framing helped a lot. How are you guys handling this right now? Are you still mostly relying on chat history, or have you moved to files as memory? Are you using one master agent file, or keeping everything repo-specific? And if you’ve found a cleaner cross-model setup than this, I’d genuinely love to hear it. If people want, I can share the exact structure. submitted by /u/kushcapital

Originally posted by u/kushcapital on r/ClaudeCode