I kept doing the same thing over and over — open Claude, ask it to act like a frontend expert, get a review, then give that review to Claude Code to implement fixes. So I automated the entire pipeline as an MCP server. Disclosure: I’m the creator of UIMax MCP. It’s free, open source (MIT), and has no paid tier. What it does: One command — “review my UI at localhost:3000” — and it:
- Captures a real screenshot of your running app (Puppeteer)
- Runs Google Lighthouse (real scores — Performance, Accessibility, Best Practices, SEO)
- Runs axe-core WCAG 2.1 accessibility audit
- Scans your source code for 25+ anti-patterns
- Returns everything to Claude Code with an expert review methodology baked in
- Claude Code generates the review AND implements every fix automatically 12 tools total, including before/after screenshot comparison, responsive viewport checks, dark mode detection, standalone HTML report export, and custom rule config. Install (one command): claude mcp add uimax – npx -y uimax-mcp@latest 100% free for Pro plan users. No API keys. No extra costs. The MCP collects the data, Claude Code (your existing subscription) does the expert analysis and fixes. Tested it on real projects — it found 109 code findings on one app, including 32 keyboard accessibility issues that would have taken 30+ minutes to find manually. GitHub: https://github.com/prembobby39-gif/uimax-mcp npm: https://npmjs.com/package/uimax-mcp Happy to answer any questions or take feedback. submitted by /u/Prem-73
Originally posted by u/Prem-73 on r/ClaudeCode
You must log in or # to comment.
