Original Reddit post

https://skarfinans.com/en/a-google-ai-breakthrough-is-pressuring-memory-chip-stocks-from-samsung-to-micron/ Google just unveiled a new compression technique called TurboQuant, and it sent memory chip stocks tumbling. The technology claims to cut the memory needed for large language models by sixfold. That is a massive reduction. Investors are worried this could slow down demand for AI memory chips. Shares of Samsung and SK Hynix fell around 5 to 6 percent in Seoul. Micron and Sandisk also took a hit in the US. A reminder of how sensitive the AI hardware market is to software breakthroughs. Anyone holding memory chip stocks right now? submitted by /u/AlphaOneYoutube

Originally posted by u/AlphaOneYoutube on r/ArtificialInteligence

  • brucethemoose@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    ·
    5 hours ago

    This is absolute nonsense on so many levels.

    • It’s just KV cache compression. It doesn’t compress the weights at all.

    • Even then, 4-bit hadamard compression is very low loss and already implemented in inference engines. This paper is better, but it wouldn’t change anything dramatically for my setup.

    • Real frontier models like Google/Deepseek-Qwen (not Grok or OpenAI) already use hybrid attention, and move father away from even needing KV cache every iteration.

    • There are hundreds of incredible papers on order-of-magnitude efficiency improvements, way bigger than this.

    It’s just investors being stupid, and YouTube hype channels doing their thing.