I’ve been using Claude Code almost since the beginning, and one recurring problem is web context. Not “can the model summarize a web page”. The harder problem is: can the agent get the right page content, in the right shape, without silently ingesting junk? For coding work, I often want the agent to do things like: read framework docs inspect changelogs compare pricing pages pull examples from docs extract structured data from a page crawl a small docs section check whether a page changed summarize a long technical article inspect brand/design info from a site So I built webclaw as an MCP server Claude Code can call directly. Available tools include: scrape crawl map batch extract summarize diff brand search research The output can be markdown, text, JSON, or structured extraction. The reason I care about MCP here is that it changes the workflow. Instead of copying a URL, opening a browser, pasting a page, trimming junk, and hoping the agent keeps the context straight, the agent can call a tool. Example workflow: “Read these docs pages and implement the SDK integration.” The agent can scrape or crawl the docs, get cleaner markdown, and use that context inside the coding task. Current state: open source Rust repo CLI MCP server hosted API 1,081 GitHub stars AGPL-3.0 works with Claude Code and other MCP-compatible clients This is not meant to replace every search/browser tool. It is more focused on the extraction layer: URL in clean context out agent uses it Question for Claude Code users: What web tasks do you actually want available as tools? Docs reading? GitHub issue analysis? Competitor pages? Changelogs? Pricing pages? Search + scrape? Something else? Repo: https://github.com/0xMassi/webclaw MCP docs: https://webclaw.io/docs/mcp submitted by /u/0xMassii
Originally posted by u/0xMassii on r/ClaudeCode
