Original Reddit post

I kept running into the same problems whenever I used Claude for coding competitions: I’d start coding before fully parsing the scoring rubric, then realize I optimized the wrong thing Context compaction mid-competition would make Claude forget key constraints My submissions lacked the polish judges notice — tests, docs, edge case handling I’d treat it like a throwaway script when winning requires product-level thinking So, I built Competitive Dominator — a Claude Skill that treats every challenge like a product launch instead of a quick hack. How it works: The skill deploys a virtual team of 13 specialized agents through a 6-phase pipeline: Intelligence Gathering — Parses the spec, extracts scoring criteria ranked by weight, identifies hidden requirements Agent Deployment — Activates the right team based on challenge type (algorithmic, ML, hackathon, CTF, LLM challenge, etc.) Architecture — Designs before coding. Complexity analysis, module structure, optimization roadmap Implementation — TDD. Tests before code. Output format validated character-by-character Optimization — Self-evaluates against scoring criteria, produces a gap analysis ranked by ROI, closes highest-value gaps first Submission — Platform-specific checklist verification. No trailing newline surprises The agents: Chief Product Manager (owns scoring rubric, kills scope creep) Solution Architect (algorithm selection, complexity analysis) Lead Developer (clean, idiomatic, documented code) Test Engineer (TDD, edge cases, fuzzing, stress tests) Code Reviewer (catches bugs before judges do) Data Scientist (activated for ML/data challenges) ML Engineer (training pipelines, LLM integration) Plus: Performance Engineer, Security Auditor, DevOps, Technical Writer, UX Designer, Risk Manager The context compaction solution: The skill maintains a CHALLENGE_STATE.md — a living document that tracks the challenge spec, every decision with reasoning, agent assignments, and progress. When Claude’s context gets compacted, it reads this file to recover full state. This was honestly the single most important feature. What’s included: 20 files, 2,450+ lines 8 agent definition files with specific responsibilities and checklists 4 reference playbooks (ML competitions, web/hackathon, challenge taxonomy, submission checklists) 2 Python scripts (state manager + self-evaluation scoring engine) — zero dependencies Works for Kaggle, Codeforces, LeetCode, hackathons, CTFs, DevPost, AI challenges Progressive disclosure — Claude only loads what’s needed for the challenge type Install: cp -r competitive-dominator ~/.claude/skills/user/competitive-dominator Also works in Claude.ai by uploading the files and telling Claude to read SKILL.md. GitHub: https://github.com/ankitjha67/competitive-dominator MIT licensed. Inspired by agency-agents, everything-claude-code, ruflo, and Karpathy’s simplicity-first philosophy. Would love feedback from anyone who’s used skills for competition workflows. What patterns have worked for you? submitted by /u/ankitjha67

Originally posted by u/ankitjha67 on r/ClaudeCode