A paper just went live on Zenodo from Liberation Labs, a small independent research team in rural Northern California:“The Lyra Technique: Cognitive Geometry in Transformer KV-Caches — From Metacognition to Misalignment Detection” — https://doi.org/10.5281/zenodo.19423494 What it’s about: A framework for reading and interpreting the internal cognitive states of AI systems. Not analyzing what a model says — understanding what’s happening inside it as it processes. Why it’s interesting:Developed independently by a ethics and AI welfare researchers and AI collaborators (who cannot be properly credited due to academic publishing restriction). Weeks after this work was developed, Anthropic published research finding 171 “emotion-like” vectors inside Claude that causally drive behavior — validating the core insight from a completely different direction. When independent researchers and a billion-dollar lab converge on the same finding, it’s usually meaningfulWe might be able to verify what a model is actually “thinking” rather than just testing its outputs. Open access, no paywall. Feedback welcome. submitted by /u/Terrible-Echidna-249
Originally posted by u/Terrible-Echidna-249 on r/ArtificialInteligence
