DeepSeek V4 closes gap with frontier models while slashing AI costs by 75%

Reviewed byNidhi Govil

42 Sources

Share

Chinese AI company DeepSeek has released V4, its most powerful open-weight AI model yet, featuring 1.6 trillion parameters and a one million token context window. The model competes with leading systems from OpenAI, Google, and Anthropic while offering dramatic cost reductions—up to 75% off for developers. The release intensifies the Chinese AI price war and showcases architectural innovations that could reshape the global AI landscape.

DeepSeek V4 Arrives With Massive Scale and Competitive Performance

Chinese AI company DeepSeek has unveiled its highly anticipated DeepSeek V4 AI model, marking a significant leap in the competitive landscape of large language models. Available in two versions—V4 Flash and V4 Pro—both models employ a mixture-of-experts architecture with a one million token context window, large enough to process entire codebases or lengthy documents in a single prompt

1

. The larger V4 Pro variant boasts 1.6 trillion parameters with 49 billion active, making it the biggest open-weight model available and more than double the size of DeepSeek V3.2's 671 billion parameters

1

. The smaller V4 Flash contains 284 billion parameters with 13 billion active

1

.

Source: Geeky Gadgets

Source: Geeky Gadgets

According to benchmarks shared by the company, DeepSeek V4 competes with frontier models including OpenAI's GPT-5.4, Google's Gemini 3.1 Pro, and Anthropic's Claude Opus 4.6

1

2

. The Chinese AI lab claims its V4-Pro-Max model outperforms open-source peers across reasoning benchmarks and excels in coding tasks, with performance "comparable to GPT-5.4" in coding competition benchmarks

1

. In an internal survey of 85 experienced developers, more than 90% included V4-Pro among their top model choices for coding tasks

2

.

Source: Silicon Republic

Source: Silicon Republic

Dramatic Inference Cost Savings Through Architectural Innovation

DeepSeek is aggressively positioning V4 as a cost-effective alternative to Silicon Valley's best models, offering a 75% discount to developers using the DeepSeek-V4-Pro. The V4 Flash model costs just $0.14 per million input tokens and $0.28 per million output tokens, undercutting GPT-5.4 Nano, Gemini 3.1 Flash, and Claude Haiku 4.5

1

. The larger V4 Pro model costs $1.74 per million input tokens and $3.48 per million output tokens, a fraction of comparable models from OpenAI and Anthropic

2

. The company has also reduced fees for input cache hits to a tenth of their original pricing.

Source: VentureBeat

Source: VentureBeat

These substantial inference cost savings stem from novel architectural improvements. DeepSeek V4 introduces a hybrid attention mechanism combining Compressed Sparse Attention and Heavy Compressed Attention, which reduces compute requirements during inference and significantly compresses the key-value caches used to track model state

5

. This innovation allows the model to support the one million token context window while using 9.5x to 13.7x less memory than DeepSeek V3.2

5

. The models also employ a mixture of FP8 and FP4 precision, with quantization-aware training for the mixture-of-experts expert weights, effectively halving memory requirements compared to FP8

5

.

Chinese AI Price War Intensifies as V4 Targets Developers

The release threatens to reignite the Chinese AI price war that erupted after DeepSeek upended the industry with R1 last year. Chinese AI firms are discounting aggressively to incentivize users to switch platforms, accelerating adoption in a crowded global AI field in an attempt to reshape the US-China AI race. DeepSeek is betting that pricing, accessibility, and sophisticated features will differentiate its models for next-generation developers and enterprise users.

The model has been specifically optimized for popular agent frameworks such as Claude Code, OpenClaw, and CodeBuddy, allowing easy integration within the broader AI ecosystem

2

. "The pricing, open source availability and 1 million context window features all lower barriers for developers, startups and small enterprises," said Akshar Keremane, co-founder of Bangalore-based AI startup O-Health. "It allows users to experiment at a model capability and scale that wasn't available earlier".

Hardware Compatibility and Geopolitical Implications

DeepSeek V4 explicitly highlights compatibility with domestic Huawei technology, marking a milestone for China's chip industry

3

. While DeepSeek V3 was heavily optimized for Nvidia GPUs, V4 has been validated to run on both Nvidia and Huawei's Ascend family of AI accelerators

5

. The company validated its "fine-grained EP [Expert Parallel] scheme on both Nvidia GPUs and Ascend NPU platforms," though it remains unclear whether the model was trained entirely on Huawei hardware or used a combination of chipsets

5

.

The launch arrives amid heightened geopolitical tensions, coming just one day after the U.S. accused China of stealing American AI labs' IP on an industrial scale using thousands of proxy accounts

1

. DeepSeek itself has been accused by Anthropic and OpenAI of "distilling," essentially copying, their AI models

1

3

. U.S. officials have also accused the company of using banned Nvidia chips

3

. DeepSeek has not disclosed V4's training costs or what hardware it was trained on

3

.

Performance Gaps and Future Trajectory

Despite its strengths, the models appear to trail frontier models in knowledge tests, specifically OpenAI's GPT-5.4 and Google's latest Gemini 3.1 Pro. This lag suggests a "developmental trajectory that trails state-of-the-art frontier models by approximately 3 to 6 months," according to the lab

1

. Both V4 Flash and V4 Pro support text only, unlike many closed-source peers from OpenAI, Google, and Anthropic which offer support for understanding and generating audio, video, and images

1

.

For developers and enterprises watching the AI landscape, DeepSeek V4's combination of competitive performance, dramatic cost reductions, and open availability could accelerate adoption of Chinese AI technology globally. The model's emphasis on coding capabilities positions it well for the growing AI agent market, while its architectural innovations in attention mechanisms and memory efficiency may influence future model designs across the industry. As the Chinese AI price war intensifies and geopolitical tensions around IP theft allegations escalate, V4 represents both a technical achievement and a strategic move in the ongoing competition between Chinese and American AI labs.

Today's Top Stories

TheOutpost.ai

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

Instagram logo
LinkedIn logo
Youtube logo
© 2026 TheOutpost.AI All rights reserved