We’ve been thinking about context wrong. Most agent frameworks treat the context window as a buffer — append until it’s full, then truncate or summarize. This works fine for chat. It’s catastrophic for procedural tasks. When an agent successfully completes a 25-step deployment — Docker builds, registry pushes, SSH connections, config edits, health checks — and then summarizes that into “deployed the app using Docker,” the knowledge is destroyed. The next time, the agent starts from scratch. Every workaround re-discovered. Every failure mode re-encountered. Every decision re-derived. SkyClaw v2.5 introduces a fundamentally different approach: the Finite Brain model. THE COGNITIVE STACK SkyClaw’s memory is now four distinct layers, each serving a different cognitive function: Skills — what the agent CAN do (tool definitions) Blueprints — what the agent KNOWS HOW to do (executable procedures) Learnings — what the agent NOTICED (ambient signals from past runs) Memory — what the agent REMEMBERS (facts, credentials, preferences) Blueprints are the core innovation. A Blueprint isn’t a summary of what happened. It’s a recipe for what to do. Exact commands. Verification steps. Failure modes and recovery paths. Decision points and what informed them. It’s the difference between a newspaper headline about surgery and an actual surgical procedure. SELF-HEALING PROCEDURES Blueprints aren’t static. They evolve through use. When a deployment procedure changes — a new migration step, a different registry endpoint, an updated config format — the Blueprint fails on first post-change execution. The agent adapts, completes the task, and refines the Blueprint. Next execution succeeds without adaptation. This is how human expertise works. A surgeon doesn’t re-learn the procedure every time. They follow a practiced sequence and refine it based on new cases. THE BRAIN SEES ITS BUDGET Every resource in SkyClaw now declares its token cost upfront. Every context rebuild includes a Resource Budget Dashboard — the agent sees exactly how much working memory it’s consumed and how much remains. When a Blueprint is too large, SkyClaw degrades gracefully: full procedure → outline only → catalog entry. Truncate before reject. Reject before crash. The system always does the best it can with the resources it has. ZERO EXTRA LLM CALLS Blueprint matching requires no dedicated LLM call. The message classifier — which already runs on every inbound message — carries a single extra field: a Blueprint category hint, picked from a grounded vocabulary of categories that actually exist in the database. Total cost: ~2ms and ~20 tokens added to an existing call. No hallucinated categories. No free-form string matching. No extra latency. The upstream call feeds the downstream decision. The context window is a finite brain. v2.5 teaches SkyClaw to think inside its skull. Github: https://github.com/nagisanzenin/skyclaw submitted by /u/No_Skill_8393
Originally posted by u/No_Skill_8393 on r/ArtificialInteligence
