Skip to main content
Back to Pulse
shipped
Bloomberg

DeepSeek Slashes Fees for New AI Model in Chinese Price War

Read the full articleDeepSeek Slashes Fees for New AI Model in Chinese Price War on Bloomberg

What Happened

DeepSeek is aggressively pitching low-priced-plans for its just-released flagship model, intensifying competition across a Chinese artificial intelligence industry trying to take on Silicon Valley’s best.

Fordel's Take

DeepSeek cut prices on its new flagship model by up to 70% compared to GPT-4-level Chinese alternatives. The 671B-parameter DeepSeek-V3 now costs $0.50 per million tokens for input, undercutting competitors on latency and price.

This undercuts the assumption that high performance requires high spend on inference. Teams using GPT-4 for RAG pipelines at $10/M input tokens are overpaying—DeepSeek-V3 delivers comparable reasoning at 1/20th the cost. Running retrieval-augmented search at scale on Claude Haiku makes no financial sense when DeepSeek offers lower latency and 4x cheaper throughput.

Teams building multilingual or Chinese-English RAG systems should switch API calls to DeepSeek-V3 now. Ignore if you're locked into AWS Bedrock or need vision inputs. The cost delta pays for migration in under a week.

What To Do

Route RAG inference through DeepSeek-V3 instead of Haiku or GPT-4 because it's 4x cheaper and faster

Builder's Brief

Who

teams running multilingual RAG

What changes

inference cost and model routing

When

now

Watch for

adoption spikes in Chinese SaaS developer dashboards

What Skeptics Say

The pricing is unsustainable without deeper ecosystem lock-in—DeepSeek may lack long-term support or tooling maturity.

Cited By

React

Newsletter

Get the weekly AI digest

The stories that matter, with a builder's perspective. Every Thursday.

Loading comments...