Original Reddit post

Here’s the prompt, new fresh windows, using sonnet on hard thinking: i have a bug in core.py: when the pipeline fails, it doesn’t restart at the checkpoint but restarts at zero: Initial run: 2705/50000 Next run: 0/50000 It should have restarted at (around) 2705 Chunks are present: ls data/.cache/test_queries/ chunk_0000.tmp chunk_0002.tmp chunk_0004.tmp chunk_0006.tmp meta.json chunk_0001.tmp chunk_0003.tmp chunk_0005.tmp chunk_0007.tmp That single prompt took 15minutes to run and burned 56% of my current session token on pro plan. I know there are hard limitations right now during peak hours. But 56% really ? For a SINGLE prompt ? The file is 362LoC (including docstrings) and it references another file that is 203LoC (also including docstrings). I’m on CLI version v2.1.90. If anyone has any idea on how to limit the token burning rate, please share. I tryed a bunch of things like reducing the the 1M context to 200k, avoid opus, clearing context regularly ect … Cheers submitted by /u/domAtOx

Originally posted by u/domAtOx on r/ClaudeCode