The algorithm that crashed Wall Street cover

The algorithm that crashed Wall Street

BuBBliK avatar

BuBBliK · @k1rallik · Mar 28

View original post

Google just crashed memory stocks with a single algorithm.

TurboQuant compresses AI memory 6x, speeds it up 8x - with zero accuracy loss. And it's free.

AI's invisible memory crisis

Every time you chat with an AI, the model doesn't just read your last message. It re-reads the entire conversation. Every token it has ever processed.

To avoid recalculating all of that from scratch, transformers store it in the KV cache - the model's short-term working memory.

The catch: this memory grows linearly with every token. A single 128K-token prompt on a 70B model eats around 40 GB of GPU memory - just for the cache. That's before the model weights even load.

At long contexts, the cache consumes more memory than the model itself.

And modern LLMs are memory-bound, not compute-bound. Generating a token is cheap math - but loading data from memory is expensive. Over and over again.

This is what limits:

  • how many users one GPU can serve
  • how long a context window the model can handle
  • how much inference costs at scale

The industry calls it the Memory Wall. And until this week, no one had a clean answer.

Google's answer: compress everything, lose nothing

Google Research published TurboQuant on March 25. It compresses the KV cache from 16 bits to 3 bits per value. 6x less memory. Zero accuracy loss.

Two stages:

PolarQuant - rotates vectors so their distribution becomes predictable. Lets you precompute the quantizer once, no calibration needed. Eliminates the 1-2 bits of overhead every previous method wastes on compression metadata.

QJL - reduces leftover error to a single sign bit. Kills bias in attention scores. Compressed output is statistically identical to full precision.

The key difference: every prior method compressed data but added overhead that ate the gains at extreme rates. TurboQuant hits near-zero overhead - approaching the Shannon limit, the theoretical floor of compression.

Plug-and-play. Works on any model. No retraining. No calibration. No fine-tuning risk.

What the benchmarks actually show

  • Memory: 6x reduction minimum. 70B model KV cache drops from ~80 GB to ~13 GB
  • Speed: up to 8x faster attention on NVIDIA H100 (4-bit vs 32-bit)
  • Accuracy: perfect score on Needle-in-a-Haystack - finding one sentence in 100,000 words
  • Cost: 50%+ cut in cloud inference spend
  • Plug-and-play on any model - Llama, Mistral, Gemma, anything

Caveat: 8x is for attention logits, not full inference. Tests only on models up to ~8B. 70B+ still unproven.

Still - a free, training-free algorithm that cuts memory 6x and hits near the theoretical compression limit. That's not incremental. That's generational.

Wall Street panicked

Investors didn't wait for peer review:

  • SanDisk: -11%
  • Micron: -7%
  • SK hynix: -6.2%
  • Samsung: -4.7%
  • NVIDIA: -4.2%
  • Philadelphia Semiconductor Index: -4.8%

All while Nasdaq was going up.

One research paper. No product. No code released. Just math - and billions wiped off memory stocks in 48 hours.

A Citrini Research analyst put it best: "It's like saying Aramco should crash because Toyota released a next-gen hybrid engine."

Wells Fargo's Andrew Rocha noted that compression algorithms have existed for years and never fundamentally changed memory procurement volumes. But this time the market didn't care about nuance - it sold first and asked questions later.

The internet figured it out immediately

If you've seen HBO's "Silicon Valley" - you already get the joke. Pied Piper, the fictional startup, built a compression algorithm that changed the rules of computing.

TurboQuant is literally the same plot. Except it's real.

The memes hit instantly. The Google Research post on X crossed 7.7 million views in under 24 hours.

Cloudflare CEO Matthew Prince called it "Google's DeepSeek moment." Someone else wrote: "Well, we all know who stole the Pied Piper codebase now." Another user calculated a Weismann Score of 5.2 - a reference only Silicon Valley fans would catch.

The comparison isn't perfect though. Pied Piper was going to change all of computing. TurboQuant only targets inference memory - not training. But as internet reactions go, the shoe fits.

The community didn't wait

Google released no official code. Just a paper with math and pseudocode.

Within 24 hours, developers built working implementations from scratch - reading formulas and writing code. This adoption speed for a research paper is almost unheard of.

What's already done:

  • Custom Triton kernel in PyTorch, tested on Gemma 3 4B on RTX 4090 - byte-identical output at 2-bit compression
  • 35B model running on Apple Silicon via MLX - 6/6 on needle-in-a-haystack at every quantization level
  • Three developers building C and CUDA implementations in llama.cpp - one reporting 18/18 tests passed
  • One person used GPT-5.4 to write a full MLX implementation in 25 minutes

One catch: an early implementer found that a naive QJL implementation produces garbage output. Without proper bias correction, quantization errors compound and the model becomes unusable. The math has to be followed exactly.

The fact that people are reimplementing a paper in hours - without official code, across Triton, MLX, and CUDA - says two things: the math is clean enough to reproduce, and the problem is urgent enough that nobody wants to wait.

The real significance isn't the compression. It's the ceiling

One of the most clear-eyed takes on TurboQuant: it matters less because it saves more memory, and more because it shows where the limit is.

Here's the compression journey so far:

  • No compression: 1x (baseline)
  • Basic quantization: 2-3x
  • Outlier-aware methods: 3-4x
  • TurboQuant in real systems: 4-4.5x

TurboQuant's error rate is already approaching the Shannon limit - the absolute theoretical floor defined by information theory. There's almost no room left to squeeze.

The paper itself proves this with a mathematical lower bound. Any quantization algorithm, no matter how clever, cannot beat this limit. TurboQuant is nearly there.

What this means: the next big breakthrough in AI efficiency won't come from compression. It will require a fundamentally different path - new architectures, new attention mechanisms, or rethinking how models store context entirely.

The plot twist: less memory needed = more memory sold

This is the Jevons Paradox. When a resource gets cheaper to use, people don't use less of it. They use more.

Save 6x on memory? Companies will run models 6x more complex. Open up use cases that were too expensive before - real-time video, million-token documents, multimodal agents running 24/7.

One GPU that served one session will now serve six. But demand won't stay at six. It will grow to sixty.

KB Securities analyst put it directly: technologies like TurboQuant lower adoption barriers and massively expand total demand. Memory makers end up as the biggest beneficiaries of AI expansion - not the victims.

The same thing happened with DeepSeek. Everyone said cheaper training would kill GPU demand. Instead it accelerated it. More people could afford to train models, so more people did.

Bottom line: memory demand will likely grow, not shrink.

Your hardware just got an upgrade. For free

No new chip. No new device. Just a software algorithm - and suddenly your existing hardware can do things it couldn't do last week.

  • Mac Mini: 100,000-token conversations with no quality loss. That's a full book-length context on a $600 machine
  • Smartphones: 32,000+ token context windows - purely through software, no hardware swap needed
  • RTX 4090: models that required multi-GPU setups now fit on a single card
  • Enterprise: cut the number of GPUs needed for long-context tasks, potentially slashing cloud spend by 50%+

The gap between local AI and cloud subscriptions just got dramatically smaller. Running serious models at home stopped being a compromise - it's becoming a real option.

The bottom line

One research paper. No code. No product launch.

48 hours later:

  • Billions wiped off memory stocks
  • Three independent implementations built from math alone
  • An entire industry recalculating demand forecasts

TurboQuant won't kill the memory market. It redrew the line between what software can solve and what still needs hardware.

And the real story? We just hit the compression ceiling. Whatever comes next - it won't be compression.

*Information sourced from:* Google Research Blog, original ICLR 2026 paper, TechCrunch, VentureBeat, Tom's Hardware, The Next Web, Investing.com, The Korea Herald, CNBC, llama.cpp GitHub discussions, turboquant.net, dejan.ai, Sketchplanations.