Original Reddit post

I’ve been running three Max accounts for 5 months across work and personal projects. The main bottleneck is context window saturation. When an agent researches, plans, codes, and reviews in a single conversation, the window fills with irrelevant data before coding begins. This causes inconsistent results. I solved this by treating the agent like a DevOps pipeline. I use isolated stages where every output is validated before moving to the next gate. I packaged this into an open-source plugin containing these skills: /research

  • Explores codebase and writes findings to a file. /plan
  • Decomposes research into issues. /pre-mortem
  • Validates the plan before coding starts. /crank
  • Spawns parallel workers with isolated contexts. /vibe
  • Validates the code, not the conversation. /post-mortem
  • Extracts learnings for next time. /rpi “goal”
  • Chains all six steps. For a quick review without the full workflow, /council performs a multi-model check on your PR. Hooks automatically enforce standards by blocking pushes without validation. A Go CLI handles knowledge injection, ensuring product personas are active at every gate. You can also define fitness goals in YAML. The /evolve command measures them, runs cycles to improve metrics, and auto-reverts regressions. Try it: npx skills@latest add boshu2/agentops --all -g Run /quickstart to begin. Works with Claude Code, Cursor, and others. Everything stays local. github.com/boshu2/agentops Feedback welcome submitted by /u/_Bo_Knows

Originally posted by u/_Bo_Knows on r/ClaudeCode