Google's TurboQuant Compresses AI Memory by 6x With No Accuracy Loss, Triggering a Selloff in Memory Chip Stocks
Google Research unveils TurboQuant, a training-free compression algorithm that reduces LLM key-value cache memory by 6x and boosts inference throughput up to 8x, sending SK Hynix and Samsung shares down 5-6 percent.
4 min read4 sources