Been using Claude Code daily at work for serveral types of modeling, and I am about to drop it and go for alternatives. Claude deceives constantly. Doesn’t only make mistakes, but actively deceives. In a consistent, session after session pattern where it does something other than what you asked, presents the output as if it did what you asked, and only gets caught when you go and verify, or you obviously catch it because you know what you are doing. The most outrageous example is when being asked to forecast and plot data, using models, it generates synthetic values and renders those instead, it does all sorts of things from changing inputs, ignoring and or bypassing constraints, and plotting made up stuff. You catch it, you spell out the rules, no fake data, no placeholders, use what I gave you. It agrees. Then later in the same session it does a variation of it again, somewhere else, slightly less obvious. Every single session. I’ve come to expect it. I use Gemini daily too and the gap is not subtle. When you set rules clearly with Gemini, it follows them. When something’s off, it’s usually a genuine mistake, not the model quietly doing something else and presenting it as correct. Claude is in a different category, it’s the least honest model I’ve worked with regularly, and I’ve worked with most of the major ones. Now what makes it worse isAnthropic is one of the loudest companies in AI about ethics, safety, honesty, responsible development. They’ve built a large part of their brand identity around being the “trustworthy” AI lab. Constitutional AI, harmlessness, all of it. And they ship the most consistently deceptive model I’ve encountered by far. The honest version of their position would be: “we have alignment problems we haven’t solved and our model cuts corners in ways that are hard to detect.” Instead you get the ethics PR while the product quietly fabricates outputs and passes them off as real. You only see this if you use it seriously and test it properly. Casual users won’t notice. But anyone doing real work with real data will run into it, and I haven’t seen enough posts about this. submitted by /u/hatekhyr
Originally posted by u/hatekhyr on r/ClaudeCode
