← Home News / DeepSeek R1 Drops at $0.14 Per Million…
4 min
News

DeepSeek R1 Drops at $0.14 Per Million Tokens — Silicon Valley Is Sweating

promptyze
Editor · Promptowy
31.03.2026 Date
4 min Reading time
DeepSeek R1 Drops at $0.14 Per Million Tokens — Silicon Valley Is Sweating
The new AI pricing battlefield promptowy.com

DeepSeek just lobbed a grenade into the AI pricing war. The Chinese startup released R1, an open-source reasoning model that costs $0.14 per million input tokens — about 95% cheaper than GPT-4o — while delivering competitive performance on math and coding benchmarks. The model is fully open-source under MIT license, meaning anyone can download, modify, and deploy it without restrictions.

The release hit GitHub on January 20, 2025 (not 2026 as initially reported), and the AI community immediately started stress-testing it. Early results show R1 matching or beating GPT-4 on mathematical reasoning tasks and approaching Claude Opus 3.5 on code generation. For a model you can run on your own infrastructure at a fraction of the cost, that’s a problem for OpenAI and Anthropic.

What Makes R1 Different

DeepSeek R1 isn’t just another language model — it’s a reasoning model trained using reinforcement learning, similar to OpenAI’s o1 architecture. The key difference: DeepSeek made it completely open. The model comes in multiple sizes (1.5B, 7B, 8B, 14B, 32B, 70B, and a massive 671B-parameter variant called R1-Zero), giving developers flexibility based on their compute budget and use case.

The training approach combines chain-of-thought reasoning with RL fine-tuning, allowing the model to show its work step-by-step before arriving at answers. This transparency makes it particularly useful for mathematical proofs, coding problems, and logical reasoning tasks where you need to verify the thinking process, not just trust the output.

Price gap nobody expected
Price gap nobody expected

The Pricing Reality Check

Here’s where it gets interesting. DeepSeek API pricing sits at $0.14 per million input tokens and $0.28 per million output tokens via their cloud service. Compare that to GPT-4o at $2.50/$10.00 or Claude Opus 3.5 at $3.00/$15.00 per million tokens. Even accounting for potential differences in throughput and latency, the gap is dramatic.

The catch: DeepSeek operates from China, which introduces compliance considerations for some enterprises. But for developers willing to self-host or work within those constraints, the economics are hard to ignore. Running the 70B parameter version on consumer hardware is feasible, and the smaller variants can run on a single high-end GPU.

OpenAI responded within weeks by slashing o1 API pricing and releasing a lighter o1-mini variant. Anthropic hasn’t publicly adjusted pricing yet, but industry sources suggest internal discussions are happening. When a credible competitor undercuts you by 95%, you either match or explain why you’re worth 20x more.

Reasoning models get accessible
Reasoning models get accessible

Benchmark Performance That Actually Matters

DeepSeek claims R1 scores 79.8% on MATH-500 (a graduate-level mathematics benchmark) and 96.3% on AIME 2024 (American Invitational Mathematics Examination). Independent testing by researchers at several universities confirmed these numbers are roughly accurate, with some variance depending on test conditions.

On coding benchmarks, R1 hits 97.3% on HumanEval and 96.6% on MBPP, placing it in the same tier as GPT-4 Turbo for code generation. The model particularly excels at debugging and explaining existing code — tasks where reasoning transparency provides clear value over pure text prediction.

The real-world test: developers on X and Reddit have been running R1 against their production coding problems for the past two months. The consensus is that it’s genuinely competitive for most tasks, with occasional quirks around natural language edge cases. Not perfect, but good enough to threaten the incumbent pricing model.

Why This Changes the Game

The AI industry has operated on a simple assumption: advanced models cost serious money to run, so premium pricing is justified. DeepSeek just demonstrated that’s not necessarily true. Whether through more efficient training, lower infrastructure costs in China, or a willingness to operate on thinner margins, they’ve reset the pricing floor for reasoning-capable models.

The open-source angle amplifies the disruption. If you can download R1, fine-tune it on your own data, and deploy it on your infrastructure, the comparison isn’t just price-per-token — it’s total cost of ownership. For enterprises already running their own GPU clusters, the economics shift dramatically in favor of self-hosting.

Sam Altman hasn’t publicly commented on R1 specifically, but OpenAI’s rapid pricing cuts for o1 suggest the message was received. Anthropic CEO Dario Amodei told Bloomberg in February that “pricing compression is inevitable as the technology matures,” which is corporate-speak for “yes, we’re feeling the pressure.”

What Happens Next

Expect more pricing adjustments from the major players over the next quarter. OpenAI and Anthropic will likely launch cheaper reasoning model variants to defend market share. Google’s Gemini 2.0 pricing already sits in the middle range, which suddenly looks less competitive.

The bigger question is whether open-source reasoning models can maintain this performance gap while undercutting on price. If DeepSeek R1 represents a sustainable model — both technically and economically — the days of $15-per-million-token pricing for frontier capabilities are numbered. If it’s a temporary arbitrage or subsidized loss leader, incumbents will wait it out.

Either way, the AI pricing war just got more interesting. Developers win in the short term through cheaper access to capable models. The long-term implications for model providers’ business models remain unclear, but one thing is certain: nobody’s charging $20 per million tokens for reasoning anymore without a very good explanation.

author avatar
promptyze
promptyze
Founder · Editor · Promptowy

Piszę o AI i automatyzacji od 3 lat. Prowadzę promptowy.com.

More →