Original Reddit post

I have a GLM subscription that’s marketed as offering 3× higher usage than Claude Pro. I primarily use it through Claude Code CLI as a backup coding model. My setup is simple: I have two Claude accounts, and when I hit usage limits on both, I switch to GLM. But honestly, I’ve been surprised by how quickly GLM gets exhausted. in practice, it seems to last less than Claude Code, despite the “3× higher usage” claim. What’s making me skeptical is the token reporting. For example, it recently showed 16 million tokens used in a single request, which feels wildly inaccurate to me. To give context: I was working on an admin panel and had already implemented 4 features using Claude Code opus before hitting the 5-hour limit. I switched to GLM 5.1 for the 5th feature, and it exhausted its usage before even finishing the task. I’ve been using GLM as a backup coding agent for around 3 months at first I thought Im overthinking but now I think something is off, and this experience makes me question whether the reported usage/token numbers are actually accurate. Has anyone else experienced something similar, or am I misunderstanding how their usage is calculated? submitted by /u/Nietzsche-og

Originally posted by u/Nietzsche-og on r/ArtificialInteligence