A lot of people are surprised that their context window gets burned through so fast. I’ve been monitoring the limits by reverse-engineering them, and you know what? Yesterday it was about a 2M context window limit for a 5-hour session, this morning it was 1.6M, and now it’s 600k — which is VERY low. That 5-hour limit is basically just a dynamic context window. By the way, the 7-day context window also jumps around, from 9M to 14M. I’m only talking about input tokens — what gets sent to the LLM — and output tokens — what the LLM generates. There’s no real point in counting cached tokens here. (PARTIALLY!) this isn’t a bug or an error — Anthropic’s answers are technically correct. But they could’ve been a bit more upfront, and then we probably wouldn’t be reacting so negatively to what they said. HOWEVER – broken consumption exacerbates this issue, making the situation significantly worse. The worst part is that not only is the 5-hour window being reduced, but the 7-day window as well—instead of 9-14 million, it’s now a 7 million window. The percentage is shrinking. (This might sound discouraging) — I recommend taking a break for now, otherwise you’ll just burn through your WEEKLY LIMITS. submitted by /u/VariousComment6946
Originally posted by u/VariousComment6946 on r/ClaudeCode
