I have a Claude Code Pro plan and have been amazed at what I have been able to produce. I have also been looking at Ralph Wiggum loops. I have a local GPU (rtx 3090 w/ 24GB vram) and have experimented with using “local” Claude ala ollama where I do (in linux) something like this: ANTHROPIC_BASE_URL=http://<mygpuserver>:11434 claude --model qwen3-coder:30b This “local” Claude seems to work ok. I do understand that my llm won’t be as capable as a frontier model like Opus, but I’m thinking I might try a Ralph loop with a free local llm. I have seen examples of creating a todo app with this, but thought I might aim a bit higher since the tokens/time are free. Does anyone have any experience with Ralph loops in local Claude, or any interesting ideas/PRDs on what to build using this process? Will this work? Any thoughts/advice? thanks submitted by /u/our_sole
Originally posted by u/our_sole on r/ClaudeCode
