I relied heavily on watching the model think and correcting it as it went along. It’s honestly what makes it possible at all for me to use this with large code bases. I frequently interrupted it to correct faulty assumptions, complement its reasoning, and steer it away from bad choices. It was really good. Now it seems like they’ve locked down and encrypted the thinking tokens. I haven’t found any official statement about it. Anyone else noticing this? It really sucks because instead of understanding what was going on, now you wait for minutes on end while it thinks and then vomits a bunch of code without any explanation. If you’ve been staring at the timer going up waiting for it to say something, you might get lucky enough to catch a mistake at that point. If you don’t, or otherwise don’t want to basically watch paint dry while it’s thinking and miss the output, you’re out of luck. Enforced vibe coding. I hate it. Anthropic is making it hard for the human to complement their product with their experience, which would be fine if AI never made a mistake. submitted by /u/gpancia
Originally posted by u/gpancia on r/ClaudeCode
