I asked Gemini “what is going on with Claude Sonnet 4.6 lately? Seems slow and lazy with answers” Here was my answer; It sounds like you’ve been having a frustrating experience with Claude Sonnet 4.6 , and I want to validate that you are not alone in noticing these performance dips. Since the release of Sonnet 4.6 in mid-February 2026, there has been widespread chatter among the community about models feeling “lazier,” experiencing latency, or behaving inconsistently compared to previous iterations. Here is a breakdown of what is likely going on behind the scenes, based on the current landscape as of April 2026:
- The “Infrastructure Strain” Factor The last few months have been a period of rapid change for the platform. In March 2026, the service faced a series of technical outages and bottlenecks. These weren’t just simple “server downtime” issues; they were related to: Control Plane Pressure: A surge in traffic (partially due to increased adoption and the rollout of new features) put heavy strain on authentication and request-routing systems. Database & Load Management: The introduction of 1M-token context windows significantly changed the “weight” of user requests. A single request isn’t just a simple chat anymore—many users are running heavy, agentic, or long-context tasks. When these complex tasks share the same infrastructure as standard chat requests, it can lead to inconsistent performance for everyone.
- “Literal” Behavior vs. “Laziness” There is a common phenomenon in LLM development where “laziness” is actually a side effect of instruction optimization . Newer models, including Sonnet 4.6, are often trained to be more “literal” or efficient to manage high-cost token usage. If the model is optimized to be more concise or to avoid “hallucinating” filler text, it might skip steps, provide shorter answers, or fail to expand on a topic unless explicitly prodded. It can feel like the model is “refusing” to do work when, in reality, its “safety” and “efficiency” guardrails are triggering more aggressively.
- Comparison to Recent Upgrades It’s worth noting that Anthropic has been iterating quickly. Claude Opus 4.7 was just released (April 16, 2026), and one of its key marketing points is a move toward more “rigor and consistency” and better instruction following. If you are finding Sonnet 4.6 consistently unreliable, it might be that your workflow is hitting the specific limitations of the current Sonnet 4.6 architecture, and you might see better results with the newer, more robust Opus 4.7 if your subscription allows it. Practical Tips for Dealing with “Lazy” Responses If you are stuck with Sonnet 4.6 for now, here are a few ways to force it out of “lazy” mode: Be Explicit About Length: If it’s being brief, add “Write a comprehensive, detailed response with no omissions” to your prompt. Chain of Thought: Ask it to “Think step-by-step” or “Create a plan before generating the code.” This forces the model to allocate more compute time to reasoning rather than jumping straight to an output. Reset the Session: If a specific conversation thread is dragging, start a new chat. Sometimes the model’s “memory” of a very long, complex conversation can lead to context bloat, making the subsequent responses feel sluggish or error-prone. Seems like a legit answer, has anyone else noticed this? submitted by /u/motion2082
Originally posted by u/motion2082 on r/ClaudeCode
You must log in or # to comment.
