3 Sources
3 Sources
[1]
Insiders Say DeepSeek V4 Will Beat Claude and ChatGPT at Coding, Launch Within Weeks - Decrypt
Developers are already hyped ahead of a potential disruption. DeepSeek is reportedly planning to drop its V4 model around mid-February, and if internal tests are any indication, Silicon Valley's AI giants should be nervous. The Hangzhou-based AI startup could be targeting a release around February 17 -- Lunar New Year, naturally -- with a model specifically engineered for coding tasks, according to The Information. People with direct knowledge of the project claim V4 outperforms both Anthropic's Claude and OpenAI's GPT series in internal benchmarks, particularly when handling extremely long code prompts. Of course, no benchmark or information about the model has been publicly shared, so it is impossible to directly verify such claims. DeepSeek hasn't confirmed the rumors either. Still, the developer community isn't waiting for official word. Reddit's r/DeepSeek and r/LocalLLaMA are already heating up, users are stockpiling API credits, and enthusiasts on X have been quick to share their predictions that V4 could cement DeepSeek's position as the scrappy underdog that refuses to play by Silicon Valley's billion-dollar rules. This wouldn't be DeepSeek's first disruption. When the company released its R1 reasoning model in January 2025, it triggered a $1 trillion sell-off in global markets. The reason? DeepSeek's R1 matched OpenAI's o1 model on math and reasoning benchmarks despite reportedly costing just $6 million to develop -- roughly 68 times cheaper than what competitors were spending. Its V3 model later hit 90.2% on the MATH-500 benchmark, blowing past Claude's 78.3% and the recent update "V3.2 Speciale" improved its performance even more. V4's coding focus would be a strategic pivot. While R1 emphasized pure reasoning -- logic, math, formal proofs -- V4 is a hybrid model (reasoning and non-reasoning tasks) that targets the enterprise developer market where high-accuracy code generation translates directly to revenue. To claim dominance, V4 would need to beat Claude Opus 4.5, which currently holds the SWE-bench Verified record at 80.9%. But if DeepSeek's past launches are any guide, then this may not be impossible to achieve even with all the constraints a Chinese AI lab would face. Assuming the rumors are true, how can this small lab achieve such a feat? The company's secret weapon could be contained in its January 1 research paper: Manifold-Constrained Hyper-Connections, or mHC. Co-authored by founder Liang Wenfeng, the new training method addresses a fundamental problem in scaling large language models -- how to expand a model's capacity without it becoming unstable or exploding during training. Traditional AI architectures force all information through a single narrow pathway. mHC widens that pathway into multiple streams that can exchange information without causing training collapse. Wei Sun, principal analyst for AI at Counterpoint Research, called mHC a "striking breakthrough" in comments to Business Insider. The technique, she said, shows DeepSeek can "bypass compute bottlenecks and unlock leaps in intelligence," even with limited access to advanced chips due to U.S. export restrictions. Lian Jye Su, chief analyst at Omdia, noted that DeepSeek's willingness to publish its methods signals a "newfound confidence in the Chinese AI industry." The company's open-source approach has made it a darling among developers who see it as embodying what OpenAI used to be, before it pivoted to closed models and billion-dollar fundraising rounds. Not everyone is convinced. Some developers on Reddit complain that DeepSeek's reasoning models waste compute on simple tasks, while critics argue the company's benchmarks don't reflect real-world messiness. One Medium post titled "DeepSeek Sucks -- And I'm Done Pretending It Doesn't" went viral in April 2025, accusing the models of producing "boilerplate nonsense with bugs" and "hallucinated libraries." DeepSeek also carries baggage. Privacy concerns have plagued the company, with some governments banning DeepSeek's native app. The company's ties to China and questions about censorship in its models add geopolitical friction to technical debates. Still, the momentum is undeniable. Deepseek has been widely adopted in Asia, and if V4 delivers on its coding promises, then enterprise adoption in the West could follow. There's also the timing. According to Reuters, DeepSeek had originally planned to release its R2 model in May 2025, but extended the runway after founder Liang became dissatisfied with its performance. Now, with V4 reportedly targeting February and R2 potentially following in August, the company is moving at a pace that suggests urgency -- or confidence. Maybe both.
[2]
DeepSeek to launch new AI model focused on coding in February: Report
Chinese AI firm DeepSeek is reportedly preparing to release its new V4 model next month. This advanced AI is expected to excel in coding tasks, possibly surpassing competitors. A key feature is its ability to process very long coding instructions. This development could significantly benefit developers working on complex software projects. Chinese AI startup DeepSeek is expected to launch its β next-generation AI model V4, featuring strong coding capabilities, in mid-February, The Information reported on Friday citing people familiar with the matter. Internal tests by DeepSeek employees suggested V4 could outperform rivals such as Anthropic's Claude and OpenAI's GPT series in coding tasks, the report said. The latest V4 model has also made breakthroughs in handling and processing extremely long coding prompts, a potential advantage for developers working on complex software projects, the Information added. Reuters could not immediately verify the report. DeepSeek did not immediately respond to a Reuters request for comment. Hangzhou-based startup DeepSeek has emerged as a key player in China's push to build its own AI ecosystem and bolster the domestic chip sector, drawing global attention after Silicon Valley β executives praised β its DeepSeek-V3 and DeepSeek-R1 models. Reuters previously reported that the Chinese artificial intelligence startup, which said in January it had built a low-cost rival to ChatGPT, has faced scrutiny in some countries over its security and privacy practices.
[3]
DeepSeek set to launch next-gen 'V4' model - The Information By Investing.com
Investing.com -- DeepSeek, the Chinese artificial intelligence upstart that rattled Silicon Valley and Wall Street last year, is preparing to launch its next-generation model in the coming weeks, according to a report from The Information. The new model, dubbed V4, is expected to feature advanced coding capabilities that internal tests suggest could leapfrog industry leaders, including OpenAI's GPT series and Anthropic's Claude. According to two people with direct knowledge of the matter cited by The Information, DeepSeek is targeting a release around the Lunar New Year in mid-February, though the timeline remains fluid. The timing of the anticipated launch follows a playbook that previously yielded massive cultural and market impact for the Beijing-based startup. Last year, DeepSeek released its flagship R1 model on January 20, just a week before China's weeklong Lunar New Year holiday. The move ensured the model dominated global tech discourse during a period of peak attention. DeepSeek, backed by the quantitative hedge fund High-Flyer Quant, became a global phenomenon following the release of R1. That "reasoning" model, designed to "think" through complex queries before answering, sent shockwaves through the AI sector not just for its performance, but for its efficiency. In a market where U.S. giants spend billions on compute, DeepSeek's ability to achieve comparable results at a fraction of the cost triggered a sharp reassessment of AI valuations and hardware dependency across Western markets. While DeepSeek's V3.2 model, released in December, reportedly outperformed OpenAI's GPT-5 and Google's Gemini 3.0 Pro on certain benchmarks, the company has yet to release a wholesale successor to its core architecture. The V4 model is positioned to fill that gap. The focus on coding is particularly significant. Programming proficiency is a primary benchmark for AI utility in enterprise environments, and a dominant V4 could further cement DeepSeek's position as a low-cost, high-performance alternative to American closed-source models. For investors, the impending release of V4 adds a new layer of volatility to the "AI arms race." When DeepSeek's R1 debuted last year, it caused a temporary rout in shares of U.S. chipmakers and AI frontrunners, as markets grappled with the reality of a Chinese player achieving parity with significantly less capital. As DeepSeek prepares to move from R1's reasoning breakthroughs to V4's coding-centric architecture, the industry is watching to see if the startup can once again disrupt the perceived dominance of San Francisco's AI titans.
Share
Share
Copy Link
Chinese AI startup DeepSeek is reportedly preparing to release its V4 model around mid-February, targeting the Lunar New Year. Internal tests suggest the next-generation AI model could outperform both Anthropic's Claude and OpenAI's GPT series in coding tasks, particularly when handling extremely long code prompts. The move could disrupt Silicon Valley's dominance in the enterprise developer market.
Chinese AI firm DeepSeek is preparing to launch its next-generation AI model V4 around mid-February, potentially coinciding with the Lunar New Year on February 17, according to a report from The Information
1
3
. People with direct knowledge of the project claim the AI model focused on coding could outperform competitors including Anthropic's Claude and OpenAI's GPT series in internal benchmarks2
. While no benchmark data has been publicly shared and DeepSeek hasn't confirmed the rumors, the developer community is already preparing for what could be another disruption from the Hangzhou-based startup.
Source: ET
Internal tests by DeepSeek employees suggest V4 could deliver coding capabilities that surpass current market leaders, particularly when handling extremely long code prompts
1
. This represents a strategic shift for the company, which previously emphasized pure reasoning with its R1 model released in January 2025. DeepSeek V4 is designed as a hybrid model capable of both reasoning and non-reasoning tasks, directly targeting the enterprise developer market where high-accuracy code generation translates to revenue1
. The ability to process very long coding instructions could prove especially valuable for developers working on complex software projects2
.The company's potential to achieve such results despite operating under U.S. export restrictions may lie in its Manifold-Constrained Hyper-Connections (mHC) training method, detailed in a January 1 research paper co-authored by founder Liang Wenfeng
1
. This technique addresses how to expand a model's capacity without causing training instability by widening information pathways into multiple streams. Wei Sun, principal analyst for AI at Counterpoint Research, called mHC a "striking breakthrough" that shows DeepSeek can "bypass compute bottlenecks and unlock leaps in intelligence" even with limited access to advanced chips1
. This low-cost, high-performance alternative approach mirrors the strategy that made DeepSeek's R1 model a phenomenon.Related Stories
DeepSeek's previous R1 release triggered a $1 trillion sell-off in global markets when it matched OpenAI's o1 model on benchmarks despite reportedly costing just $6 million to developβroughly 68 times cheaper than competitors
1
. The startup, backed by quantitative hedge fund High-Flyer Quant, has emerged as a key player in China's push to build its own AI ecosystem2
3
. To claim dominance in coding, V4 would need to beat Claude Opus 4.5, which currently holds the SWE-bench Verified record at 80.9%1
. For investors, the impending V4 release adds volatility to the AI arms race, as markets reassess hardware dependency and market valuations across Western AI companies3
.The developer community isn't waiting for official confirmation, with Reddit's r/DeepSeek and r/LocalLLaMA heating up as users stockpile API credits
1
. Lian Jye Su, chief analyst at Omdia, noted that DeepSeek's willingness to publish its methods signals "newfound confidence in the Chinese AI industry," with the company's open-source approach making it popular among developers who see it as embodying what OpenAI used to be1
. However, privacy concerns have plagued the company, with some governments banning DeepSeek's native app1
2
. Critics also argue the company's benchmarks don't reflect real-world complexity, with some developers complaining about wasted compute on simple tasks1
. Still, if V4 delivers on its coding promises, enterprise adoption in Western markets could follow the model's widespread use across Asia.Summarized by
Navi
[1]
1
Policy and Regulation

2
Technology

3
Policy and Regulation
