Skip to main content
Back to Pulse
Hugging Face

Large Language Models: A New Moore’s Law?

Read the full articleLarge Language Models: A New Moore’s Law? on Hugging Face

What Happened

Large Language Models: A New Moore’s Law?

Fordel's Take

Since GPT-3, LLM capability-per-dollar has roughly doubled every 8–12 months. GPT-4o now costs 1/20th what GPT-4 cost at launch for equivalent token throughput. No other infrastructure category has deflated this fast.

Teams designing RAG pipelines today are hardcoding cost assumptions that will be wrong within two quarters. Locking every agent call to Claude Opus because it topped your 2024 evals is just burning budget. Claude Haiku or GPT-4o mini handles 80% of classification and extraction tasks at 50x lower cost.

What To Do

Route classification and extraction calls to Claude Haiku instead of Opus because the capability gap is negligible and the cost gap is 50x.

Cited By

React

Newsletter

Get the weekly AI digest

The stories that matter, with a builder's perspective. Every Thursday.

Loading comments...