Original Reddit post

We all know the problem: the longer you chat with an AI, the “dumber” it seems to get. Context windows are getting bigger, but the “lost in the middle” effect is still very real. Plus, dragging 100k tokens of history into every new message is a massive waste of energy and tokens. I spent the last few weeks building Lakon , a Continuity Engine designed to extract the “State” from a conversation. How it works technically: It doesn’t just summarize. It uses a Map-Reduce pipeline to chunk your history and systematically extract goals, decisions, and hard constraints. It then generates a structured briefing that brings a new AI instance fully up to speed in seconds. It also includes a browser extension that does Semantic Compression

stripping away hedging, politeness, and redundant phrasing in real-time to shrink your daily prompts by ~70% without losing intent. It’s free, zero-retention (I don’t store your logs), and the code is on GitHub. Try it out: Lakon Web GitHub: Lakon-Public Would love any feedback on the extraction logic and the product! submitted by /u/PriorNervous1031

Originally posted by u/PriorNervous1031 on r/ArtificialInteligence