China

DeepSeek V4 Returns to Challenge US AI Giants With Aggressive Pricing

DeepSeek. © Solen Feyissa auf Unsplash
DeepSeek. © Solen Feyissa auf Unsplash

DeepSeek achieved worldwide fame in 2025, sending shockwaves through the US-dominated AI industry in particular. At the time, the cost-efficient open-source model V3 was released, which was on par with US models while being significantly cheaper. Since then, a lot has happened, especially in China, and DeepSeek has not been the number one in open-source LLMs for some time; startups such as Z.ai (GLM-5) or Moonshot AI (Kimi K2.6) have overtaken it.

Now, however, DeepSeek is fighting its way back to the front with the release of V4. The two models DeepSeek-V4-Pro and DeepSeek-V4-Flash are available immediately and are being released simultaneously as open-source models. Particularly noteworthy is the combination of high performance and significantly lower costs compared to the leading American competitor models. Currently, V4-Pro ranks just behind GLM-5 in the open-source category on Arena.ai, but only manages 20th place in the overall ranking of all LLMs.

Two Models for Different Requirements

The V4 series consists of two variants that differ in size and performance profile. Both models support a context window of one million tokens and are built on a Mixture-of-Experts architecture. They support both a normal mode and a thinking mode, in which the thinking intensity can be adjusted via the reasoning_effort parameter.

  • DeepSeek-V4-Pro has a total of 1.6 trillion parameters, of which 49 billion are active. According to DeepSeek, the model achieves performance comparable to leading closed models, particularly in the areas of mathematics, STEM, and coding competitions.
  • DeepSeek-V4-Flash has a total of 284 billion parameters, of which 13 billion are active. It is faster and more cost-effective, showing similar performance to the Pro model on simple tasks, but falling behind on complex tasks.

Performance on Par with the World’s Best

According to its own statements, DeepSeek-V4-Pro outperforms all currently publicly evaluated open-source models in the areas of mathematics, STEM, and coding competitions. In terms of world knowledge, the model falls just short of the leading closed model Gemini-Pro-3.1, but clearly outperforms all other open-source alternatives.

The improved agent capabilities are particularly highlighted. In internal tests, the user experience was rated as better than Claude Sonnet 4.5, and the output quality is said to approach that of Opus 4.6 in normal mode. According to DeepSeek, V4-Pro has already become the internal default model for coding tasks within the company.

Technically, DeepSeek-V4 employs a novel attention mechanism that compresses tokens and is combined with the proprietary DeepSeek Sparse Attention. This enables the long context while significantly reducing computational and memory requirements compared to conventional methods.

The Decisive Advantage: The Prices

What makes DeepSeek V4 particularly interesting is the aggressive pricing. Both models significantly undercut their direct competitors. DeepSeek-V4-Flash is the cheapest model in its class and even beats OpenAI’s GPT-5.4 Nano. DeepSeek-V4-Pro, in turn, is the most affordable of the high-performance frontier models and costs only a fraction compared to Claude Sonnet 4.6 or GPT-5.4.

Model Input ($/M tokens) Output ($/M tokens)
DeepSeek V4 Flash $0.14 $0.28
GPT-5.4 Nano $0.20 $1.25
Gemini 3.1 Flash-Lite $0.25 $1.50
Gemini 3 Flash Preview $0.50 $3.00
GPT-5.4 Mini $0.75 $4.50
Claude Haiku 4.5 $1.00 $5.00
DeepSeek V4 Pro $1.74 $3.48
Gemini 3.1 Pro $2.00 $12.00
GPT-5.4 $2.50 $15.00
Claude Sonnet 4.6 $3.00 $15.00
Claude Opus 4.7 $5.00 $25.00
GPT-5.5 $5.00 $30.00

The price differences are substantial. Those who use DeepSeek-V4-Pro instead of Claude Sonnet 4.6 pay more than four times less for output. Compared to GPT-5.5 or Claude Opus 4.7, the difference is even more drastic. According to DeepSeek, this is made possible by the high efficiency of the new architecture, which significantly reduces computational and memory requirements, especially for long contexts.

Availability and API Access

Both models are accessible via the official website chat.deepseek.com as well as through the DeepSeek app. The API supports the interfaces of OpenAI ChatCompletions and Anthropic. To access the new models, only the model name needs to be changed to deepseek-v4-pro or deepseek-v4-flash. The previous model names deepseek-chat and deepseek-reasoner will be discontinued in three months, specifically on July 24, 2026.

Rank My Startup: Erobere die Liga der Top Founder!
Advertisement
Advertisement

Specials from our Partners

Top Posts from our Network

Deep Dives

© Wiener Börse

IPO Spotlight

powered by Wiener Börse

Europe's Top Unicorn Investments 2023

The full list of companies that reached a valuation of € 1B+ this year
© Behnam Norouzi on Unsplash

Crypto Investment Tracker 2022

The biggest deals in the industry, ranked by Trending Topics
ThisisEngineering RAEng on Unsplash

Technology explained

Powered by PwC
© addendum

Inside the Blockchain

Die revolutionäre Technologie von Experten erklärt

Trending Topics Tech Talk

Der Podcast mit smarten Köpfen für smarte Köpfe
© Shannon Rowies on Unsplash

We ❤️ Founders

Die spannendsten Persönlichkeiten der Startup-Szene
Tokio bei Nacht und Regen. © Unsplash

🤖Big in Japan🤖

Startups - Robots - Entrepreneurs - Tech - Trends

Continue Reading