Original Reddit post

We keep coming across claims of Opus being lobotomised every once in a while, and them being validated by some and debunked by others. It is just sad that AI companies can now get away with gaslighting us given how difficult it can be to establish whether a model is genuinely performing better or worse than it’s baseline. Codex review workflow, however, has made it possible for me to pick up when Opus goes dumb. When it starts to pick up P0 gaps in Opus’s implementations, even though you ran a session in maximum thinking mode, and used plan tool, and /simplify skill after… Incredibly stupid oversights keep floating up these days. It’s just shameful that a 20$ Codex is steering my 200$ Opus around like this. And at least I feel sure that I am not imagining it: Opus HAS gone dumb in the past week or two . And that is ON TOP of it going slow. It’s just sad. I hope open weight models catch up. Lack of competition in the high-tier AI models market is blowing up. submitted by /u/ThePurpleAbsurdist

Originally posted by u/ThePurpleAbsurdist on r/ClaudeCode