Generating code is a solved issue. But keeping the product from derailing is still a struggle. We need to set up some kind of feedback loop that tells the agent what is working and what needs fixing. While agents can generate test automation, most of this feedback loop still involves human labor. But for how long? I’m running an experiment where an agent builds a Doom clone overnight and I give feedback if it needs steering. If there is no human feedback, the agent makes up new features. The goal is to see how long we can keep this running until a human needs to intervene. The first nights were rocky, but now the loop is operational. The game is playable and there is a daily blog of the new updates. Check out Daily Doom. Or read the related blog post. submitted by /u/MikeNonect
Originally posted by u/MikeNonect on r/ClaudeCode
