Skip to main content
Back to Pulse
Cloudflare

Unweight: how we compressed an LLM 22% without sacrificing quality

Read the full articleUnweight: how we compressed an LLM 22% without sacrificing quality on Cloudflare

What Happened

Running LLMs across Cloudflare’s network requires us to be smarter and more efficient about GPU memory bandwidth. That’s why we developed Unweight, a lossless inference-time compression system that achieves up to a 22% model footprint reduction, so that we can deliver faster and cheaper inference th

Cited By

React

Newsletter

Get the weekly AI digest

The stories that matter, with a builder's perspective. Every Thursday.

Loading comments...